Read the Docs build information Build id: 2698162 Project: nvidia-nemo Version: 11450 Commit: 62d413f68b0c4cb8b7425ca60df050aca182a1fc Date: 2024-12-19T21:09:16.709432Z State: finished Success: True [rtd-command-info] start-time: 2024-12-19T21:09:17.860198Z, end-time: 2024-12-19T21:09:23.281104Z, duration: 5, exit-code: 0 git clone --depth 1 https://github.com/NVIDIA/NeMo.git . Cloning into '.'... [rtd-command-info] start-time: 2024-12-19T21:09:23.344658Z, end-time: 2024-12-19T21:09:24.710042Z, duration: 1, exit-code: 0 git fetch origin --force --prune --prune-tags --depth 50 pull/11450/head:external-11450 From https://github.com/NVIDIA/NeMo * [new ref] refs/pull/11450/head -> external-11450 [rtd-command-info] start-time: 2024-12-19T21:09:24.829645Z, end-time: 2024-12-19T21:09:25.315998Z, duration: 0, exit-code: 0 git checkout --force 62d413f68b0c4cb8b7425ca60df050aca182a1fc Note: switching to '62d413f68b0c4cb8b7425ca60df050aca182a1fc'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by switching back to a branch. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at 62d413f Merge branch 'speechllm-develop-gen-duplex' into speechllm-develop-gen_duplex2_clean [rtd-command-info] start-time: 2024-12-19T21:09:25.387514Z, end-time: 2024-12-19T21:09:25.446848Z, duration: 0, exit-code: 0 cat .readthedocs.yml # ============================================================================= # Copyright (c) 2020 NVIDIA. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================= # Read the Docs configuration file # See https://docs.readthedocs.io/en/stable/config-file/v2.html for details # Required field. version: 2 build: os: ubuntu-22.04 tools: python: "3.10" # Build documentation in the docs/ directory with Sphinx. sphinx: configuration: docs/source/conf.py # Set the version of Python and requirements required to build your docs python: install: - requirements: requirements/requirements_docs.txt [rtd-command-info] start-time: 2024-12-19T21:09:30.573000Z, end-time: 2024-12-19T21:09:30.652837Z, duration: 0, exit-code: 0 asdf global python 3.10.15 [rtd-command-info] start-time: 2024-12-19T21:09:31.064236Z, end-time: 2024-12-19T21:09:32.558185Z, duration: 1, exit-code: 0 python -mvirtualenv $READTHEDOCS_VIRTUALENV_PATH created virtual environment CPython3.10.15.final.0-64 in 929ms creator CPython3Posix(dest=/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450, clear=False, no_vcs_ignore=False, global=False) seeder FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=/home/docs/.local/share/virtualenv) added seed packages: pip==23.1, setuptools==67.6.1, wheel==0.40.0 activators BashActivator,CShellActivator,FishActivator,NushellActivator,PowerShellActivator,PythonActivator [rtd-command-info] start-time: 2024-12-19T21:09:32.630999Z, end-time: 2024-12-19T21:09:44.164639Z, duration: 11, exit-code: 0 python -m pip install --upgrade --no-cache-dir pip setuptools Requirement already satisfied: pip in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (23.1) Collecting pip Downloading pip-24.3.1-py3-none-any.whl (1.8 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.8/1.8 MB 8.8 MB/s eta 0:00:00 Requirement already satisfied: setuptools in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (67.6.1) Collecting setuptools Downloading setuptools-75.6.0-py3-none-any.whl (1.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 19.8 MB/s eta 0:00:00 Installing collected packages: setuptools, pip Attempting uninstall: setuptools Found existing installation: setuptools 67.6.1 Uninstalling setuptools-67.6.1: Successfully uninstalled setuptools-67.6.1 Attempting uninstall: pip Found existing installation: pip 23.1 Uninstalling pip-23.1: Successfully uninstalled pip-23.1 Successfully installed pip-24.3.1 setuptools-75.6.0 [rtd-command-info] start-time: 2024-12-19T21:09:44.229789Z, end-time: 2024-12-19T21:09:50.350787Z, duration: 6, exit-code: 0 python -m pip install --upgrade --no-cache-dir sphinx Collecting sphinx Downloading sphinx-8.1.3-py3-none-any.whl.metadata (6.4 kB) Collecting sphinxcontrib-applehelp>=1.0.7 (from sphinx) Downloading sphinxcontrib_applehelp-2.0.0-py3-none-any.whl.metadata (2.3 kB) Collecting sphinxcontrib-devhelp>=1.0.6 (from sphinx) Downloading sphinxcontrib_devhelp-2.0.0-py3-none-any.whl.metadata (2.3 kB) Collecting sphinxcontrib-htmlhelp>=2.0.6 (from sphinx) Downloading sphinxcontrib_htmlhelp-2.1.0-py3-none-any.whl.metadata (2.3 kB) Collecting sphinxcontrib-jsmath>=1.0.1 (from sphinx) Downloading sphinxcontrib_jsmath-1.0.1-py2.py3-none-any.whl.metadata (1.4 kB) Collecting sphinxcontrib-qthelp>=1.0.6 (from sphinx) Downloading sphinxcontrib_qthelp-2.0.0-py3-none-any.whl.metadata (2.3 kB) Collecting sphinxcontrib-serializinghtml>=1.1.9 (from sphinx) Downloading sphinxcontrib_serializinghtml-2.0.0-py3-none-any.whl.metadata (2.4 kB) Collecting Jinja2>=3.1 (from sphinx) Downloading jinja2-3.1.4-py3-none-any.whl.metadata (2.6 kB) Collecting Pygments>=2.17 (from sphinx) Downloading pygments-2.18.0-py3-none-any.whl.metadata (2.5 kB) Collecting docutils<0.22,>=0.20 (from sphinx) Downloading docutils-0.21.2-py3-none-any.whl.metadata (2.8 kB) Collecting snowballstemmer>=2.2 (from sphinx) Downloading snowballstemmer-2.2.0-py2.py3-none-any.whl.metadata (6.5 kB) Collecting babel>=2.13 (from sphinx) Downloading babel-2.16.0-py3-none-any.whl.metadata (1.5 kB) Collecting alabaster>=0.7.14 (from sphinx) Downloading alabaster-1.0.0-py3-none-any.whl.metadata (2.8 kB) Collecting imagesize>=1.3 (from sphinx) Downloading imagesize-1.4.1-py2.py3-none-any.whl.metadata (1.5 kB) Collecting requests>=2.30.0 (from sphinx) Downloading requests-2.32.3-py3-none-any.whl.metadata (4.6 kB) Collecting packaging>=23.0 (from sphinx) Downloading packaging-24.2-py3-none-any.whl.metadata (3.2 kB) Collecting tomli>=2 (from sphinx) Downloading tomli-2.2.1-py3-none-any.whl.metadata (10 kB) Collecting MarkupSafe>=2.0 (from Jinja2>=3.1->sphinx) Downloading MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.0 kB) Collecting charset-normalizer<4,>=2 (from requests>=2.30.0->sphinx) Downloading charset_normalizer-3.4.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (34 kB) Collecting idna<4,>=2.5 (from requests>=2.30.0->sphinx) Downloading idna-3.10-py3-none-any.whl.metadata (10 kB) Collecting urllib3<3,>=1.21.1 (from requests>=2.30.0->sphinx) Downloading urllib3-2.2.3-py3-none-any.whl.metadata (6.5 kB) Collecting certifi>=2017.4.17 (from requests>=2.30.0->sphinx) Downloading certifi-2024.12.14-py3-none-any.whl.metadata (2.3 kB) Downloading sphinx-8.1.3-py3-none-any.whl (3.5 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.5/3.5 MB 135.2 MB/s eta 0:00:00 Downloading alabaster-1.0.0-py3-none-any.whl (13 kB) Downloading babel-2.16.0-py3-none-any.whl (9.6 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 9.6/9.6 MB 242.2 MB/s eta 0:00:00 Downloading docutils-0.21.2-py3-none-any.whl (587 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 587.4/587.4 kB 502.5 MB/s eta 0:00:00 Downloading imagesize-1.4.1-py2.py3-none-any.whl (8.8 kB) Downloading jinja2-3.1.4-py3-none-any.whl (133 kB) Downloading packaging-24.2-py3-none-any.whl (65 kB) Downloading pygments-2.18.0-py3-none-any.whl (1.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 596.3 MB/s eta 0:00:00 Downloading requests-2.32.3-py3-none-any.whl (64 kB) Downloading snowballstemmer-2.2.0-py2.py3-none-any.whl (93 kB) Downloading sphinxcontrib_applehelp-2.0.0-py3-none-any.whl (119 kB) Downloading sphinxcontrib_devhelp-2.0.0-py3-none-any.whl (82 kB) Downloading sphinxcontrib_htmlhelp-2.1.0-py3-none-any.whl (98 kB) Downloading sphinxcontrib_jsmath-1.0.1-py2.py3-none-any.whl (5.1 kB) Downloading sphinxcontrib_qthelp-2.0.0-py3-none-any.whl (88 kB) Downloading sphinxcontrib_serializinghtml-2.0.0-py3-none-any.whl (92 kB) Downloading tomli-2.2.1-py3-none-any.whl (14 kB) Downloading certifi-2024.12.14-py3-none-any.whl (164 kB) Downloading charset_normalizer-3.4.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (144 kB) Downloading idna-3.10-py3-none-any.whl (70 kB) Downloading MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (20 kB) Downloading urllib3-2.2.3-py3-none-any.whl (126 kB) Installing collected packages: snowballstemmer, urllib3, tomli, sphinxcontrib-serializinghtml, sphinxcontrib-qthelp, sphinxcontrib-jsmath, sphinxcontrib-htmlhelp, sphinxcontrib-devhelp, sphinxcontrib-applehelp, Pygments, packaging, MarkupSafe, imagesize, idna, docutils, charset-normalizer, certifi, babel, alabaster, requests, Jinja2, sphinx Successfully installed Jinja2-3.1.4 MarkupSafe-3.0.2 Pygments-2.18.0 alabaster-1.0.0 babel-2.16.0 certifi-2024.12.14 charset-normalizer-3.4.0 docutils-0.21.2 idna-3.10 imagesize-1.4.1 packaging-24.2 requests-2.32.3 snowballstemmer-2.2.0 sphinx-8.1.3 sphinxcontrib-applehelp-2.0.0 sphinxcontrib-devhelp-2.0.0 sphinxcontrib-htmlhelp-2.1.0 sphinxcontrib-jsmath-1.0.1 sphinxcontrib-qthelp-2.0.0 sphinxcontrib-serializinghtml-2.0.0 tomli-2.2.1 urllib3-2.2.3 [rtd-command-info] start-time: 2024-12-19T21:09:50.439099Z, end-time: 2024-12-19T21:09:59.455217Z, duration: 9, exit-code: 0 python -m pip install --exists-action=w --no-cache-dir -r requirements/requirements_docs.txt Collecting boto3 (from -r requirements/requirements_docs.txt (line 1)) Downloading boto3-1.35.85-py3-none-any.whl.metadata (6.7 kB) Requirement already satisfied: Jinja2 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from -r requirements/requirements_docs.txt (line 2)) (3.1.4) Collecting latexcodec (from -r requirements/requirements_docs.txt (line 3)) Downloading latexcodec-3.0.0-py3-none-any.whl.metadata (4.9 kB) Collecting numpy (from -r requirements/requirements_docs.txt (line 4)) Downloading numpy-2.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (62 kB) Collecting pydata-sphinx-theme (from -r requirements/requirements_docs.txt (line 5)) Downloading pydata_sphinx_theme-0.16.1-py3-none-any.whl.metadata (7.5 kB) Requirement already satisfied: Sphinx in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from -r requirements/requirements_docs.txt (line 6)) (8.1.3) Collecting sphinx-book-theme (from -r requirements/requirements_docs.txt (line 7)) Downloading sphinx_book_theme-1.1.3-py3-none-any.whl.metadata (5.7 kB) Collecting sphinx-copybutton (from -r requirements/requirements_docs.txt (line 8)) Downloading sphinx_copybutton-0.5.2-py3-none-any.whl.metadata (3.2 kB) Collecting sphinxcontrib-bibtex (from -r requirements/requirements_docs.txt (line 9)) Downloading sphinxcontrib_bibtex-2.6.3-py3-none-any.whl.metadata (6.3 kB) Collecting sphinxext-opengraph (from -r requirements/requirements_docs.txt (line 10)) Downloading sphinxext_opengraph-0.9.1-py3-none-any.whl.metadata (6.5 kB) Requirement already satisfied: urllib3 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from -r requirements/requirements_docs.txt (line 11)) (2.2.3) Collecting wrapt (from -r requirements/requirements_docs.txt (line 12)) Downloading wrapt-1.17.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.4 kB) Collecting botocore<1.36.0,>=1.35.85 (from boto3->-r requirements/requirements_docs.txt (line 1)) Downloading botocore-1.35.85-py3-none-any.whl.metadata (5.7 kB) Collecting jmespath<2.0.0,>=0.7.1 (from boto3->-r requirements/requirements_docs.txt (line 1)) Downloading jmespath-1.0.1-py3-none-any.whl.metadata (7.6 kB) Collecting s3transfer<0.11.0,>=0.10.0 (from boto3->-r requirements/requirements_docs.txt (line 1)) Downloading s3transfer-0.10.4-py3-none-any.whl.metadata (1.7 kB) Requirement already satisfied: MarkupSafe>=2.0 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Jinja2->-r requirements/requirements_docs.txt (line 2)) (3.0.2) Collecting beautifulsoup4 (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) Downloading beautifulsoup4-4.12.3-py3-none-any.whl.metadata (3.8 kB) Requirement already satisfied: docutils!=0.17.0 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) (0.21.2) Requirement already satisfied: Babel in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) (2.16.0) Requirement already satisfied: pygments>=2.7 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) (2.18.0) Collecting accessible-pygments (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) Downloading accessible_pygments-0.0.5-py3-none-any.whl.metadata (10 kB) Collecting typing-extensions (from pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) Downloading typing_extensions-4.12.2-py3-none-any.whl.metadata (3.0 kB) Requirement already satisfied: sphinxcontrib-applehelp>=1.0.7 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.0.0) Requirement already satisfied: sphinxcontrib-devhelp>=1.0.6 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.0.0) Requirement already satisfied: sphinxcontrib-htmlhelp>=2.0.6 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.1.0) Requirement already satisfied: sphinxcontrib-jsmath>=1.0.1 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (1.0.1) Requirement already satisfied: sphinxcontrib-qthelp>=1.0.6 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.0.0) Requirement already satisfied: sphinxcontrib-serializinghtml>=1.1.9 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.0.0) Requirement already satisfied: snowballstemmer>=2.2 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.2.0) Requirement already satisfied: alabaster>=0.7.14 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (1.0.0) Requirement already satisfied: imagesize>=1.3 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (1.4.1) Requirement already satisfied: requests>=2.30.0 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.32.3) Requirement already satisfied: packaging>=23.0 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (24.2) Requirement already satisfied: tomli>=2 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from Sphinx->-r requirements/requirements_docs.txt (line 6)) (2.2.1) Collecting pybtex>=0.24 (from sphinxcontrib-bibtex->-r requirements/requirements_docs.txt (line 9)) Downloading pybtex-0.24.0-py2.py3-none-any.whl.metadata (2.0 kB) Collecting pybtex-docutils>=1.0.0 (from sphinxcontrib-bibtex->-r requirements/requirements_docs.txt (line 9)) Downloading pybtex_docutils-1.0.3-py3-none-any.whl.metadata (4.3 kB) Collecting python-dateutil<3.0.0,>=2.1 (from botocore<1.36.0,>=1.35.85->boto3->-r requirements/requirements_docs.txt (line 1)) Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl.metadata (8.4 kB) Collecting PyYAML>=3.01 (from pybtex>=0.24->sphinxcontrib-bibtex->-r requirements/requirements_docs.txt (line 9)) Downloading PyYAML-6.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB) Collecting six (from pybtex>=0.24->sphinxcontrib-bibtex->-r requirements/requirements_docs.txt (line 9)) Downloading six-1.17.0-py2.py3-none-any.whl.metadata (1.7 kB) Requirement already satisfied: charset-normalizer<4,>=2 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from requests>=2.30.0->Sphinx->-r requirements/requirements_docs.txt (line 6)) (3.4.0) Requirement already satisfied: idna<4,>=2.5 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from requests>=2.30.0->Sphinx->-r requirements/requirements_docs.txt (line 6)) (3.10) Requirement already satisfied: certifi>=2017.4.17 in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages (from requests>=2.30.0->Sphinx->-r requirements/requirements_docs.txt (line 6)) (2024.12.14) Collecting soupsieve>1.2 (from beautifulsoup4->pydata-sphinx-theme->-r requirements/requirements_docs.txt (line 5)) Downloading soupsieve-2.6-py3-none-any.whl.metadata (4.6 kB) Downloading boto3-1.35.85-py3-none-any.whl (139 kB) Downloading latexcodec-3.0.0-py3-none-any.whl (18 kB) Downloading numpy-2.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (16.4 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 16.4/16.4 MB 176.6 MB/s eta 0:00:00 Downloading pydata_sphinx_theme-0.16.1-py3-none-any.whl (6.7 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.7/6.7 MB 208.9 MB/s eta 0:00:00 Downloading sphinx_book_theme-1.1.3-py3-none-any.whl (430 kB) Downloading sphinx_copybutton-0.5.2-py3-none-any.whl (13 kB) Downloading sphinxcontrib_bibtex-2.6.3-py3-none-any.whl (40 kB) Downloading sphinxext_opengraph-0.9.1-py3-none-any.whl (1.0 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.0/1.0 MB 595.4 MB/s eta 0:00:00 Downloading wrapt-1.17.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (82 kB) Downloading botocore-1.35.85-py3-none-any.whl (13.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13.3/13.3 MB 204.0 MB/s eta 0:00:00 Downloading jmespath-1.0.1-py3-none-any.whl (20 kB) Downloading pybtex-0.24.0-py2.py3-none-any.whl (561 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 561.4/561.4 kB 556.4 MB/s eta 0:00:00 Downloading pybtex_docutils-1.0.3-py3-none-any.whl (6.4 kB) Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) Downloading accessible_pygments-0.0.5-py3-none-any.whl (1.4 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 567.9 MB/s eta 0:00:00 Downloading beautifulsoup4-4.12.3-py3-none-any.whl (147 kB) Downloading typing_extensions-4.12.2-py3-none-any.whl (37 kB) Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB) Downloading PyYAML-6.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (751 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 751.2/751.2 kB 539.8 MB/s eta 0:00:00 Downloading six-1.17.0-py2.py3-none-any.whl (11 kB) Downloading soupsieve-2.6-py3-none-any.whl (36 kB) Installing collected packages: wrapt, typing-extensions, soupsieve, six, PyYAML, numpy, latexcodec, jmespath, accessible-pygments, python-dateutil, pybtex, beautifulsoup4, sphinxext-opengraph, sphinx-copybutton, pydata-sphinx-theme, pybtex-docutils, botocore, sphinxcontrib-bibtex, sphinx-book-theme, s3transfer, boto3 Successfully installed PyYAML-6.0.2 accessible-pygments-0.0.5 beautifulsoup4-4.12.3 boto3-1.35.85 botocore-1.35.85 jmespath-1.0.1 latexcodec-3.0.0 numpy-2.2.0 pybtex-0.24.0 pybtex-docutils-1.0.3 pydata-sphinx-theme-0.16.1 python-dateutil-2.9.0.post0 s3transfer-0.10.4 six-1.17.0 soupsieve-2.6 sphinx-book-theme-1.1.3 sphinx-copybutton-0.5.2 sphinxcontrib-bibtex-2.6.3 sphinxext-opengraph-0.9.1 typing-extensions-4.12.2 wrapt-1.17.0 [rtd-command-info] start-time: 2024-12-19T21:09:59.614072Z, end-time: 2024-12-19T21:09:59.678935Z, duration: 0, exit-code: 0 cat docs/source/conf.py #!/usr/bin/env python3 # -*- coding: utf-8 -*- # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import re import sys import glob import sphinx_book_theme # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. sys.path.insert(0, os.path.abspath("../..")) sys.path.insert(0, os.path.abspath("../../nemo")) from package_info import __version__ templates_path = ["_templates"] autodoc_mock_imports = [ 'torch', 'torch.nn', 'torch.utils', 'torch.optim', 'torch.utils.data', 'torch.utils.data.sampler', 'torchtext', 'torchvision', 'ruamel.yaml', # ruamel.yaml has ., which is troublesome for this regex 'hydra', # hydra-core in requirements, hydra during import 'dateutil', # part of core python 'transformers.tokenization_bert', # has ., troublesome for this regex 'sklearn', # scikit_learn in requirements, sklearn in import 'nemo_text_processing.inverse_text_normalization', # Not installed automatically 'nemo_text_processing.text_normalization', # Not installed automatically 'attr', # attrdict in requirements, attr in import 'torchmetrics', # inherited from PTL 'lightning_utilities', # inherited from PTL 'lightning_fabric', 'apex', 'megatron.core', 'transformer_engine', 'joblib', # inherited from optional code 'IPython', 'ipadic', 'psutil', 'regex', 'PIL', 'boto3', 'taming', 'cytoolz', # for adapters 'megatron', # for nlp ] _skipped_autodoc_mock_imports = ['wrapt', 'numpy'] for req_path in sorted(list(glob.glob("../../requirements/*.txt"))): if "docs.txt" in req_path: continue req_file = os.path.abspath(os.path.expanduser(req_path)) with open(req_file, 'r') as f: for line in f: line = line.replace("\n", "") req = re.search(r"([a-zA-Z0-9-_]*)", line) if req: req = req.group(1) req = req.replace("-", "_") if req not in autodoc_mock_imports: if req in _skipped_autodoc_mock_imports: print(f"Skipping req : `{req}` (lib {line})") continue autodoc_mock_imports.append(req) print(f"Adding req : `{req}` to autodoc mock requirements (lib {line})") else: print(f"`{req}` already added to autodoc mock requirements (lib {line})") # # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. # # needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ "sphinx.ext.autodoc", "sphinx.ext.todo", "sphinx.ext.coverage", "sphinx.ext.mathjax", "sphinx.ext.ifconfig", "sphinx.ext.viewcode", "sphinx.ext.napoleon", "sphinx.ext.githubpages", "sphinx.ext.inheritance_diagram", "sphinx.ext.intersphinx", # "sphinx.ext.autosectionlabel", "sphinxcontrib.bibtex", "sphinx_copybutton", "sphinxext.opengraph", ] bibtex_bibfiles = [ 'asr/asr_all.bib', 'nlp/nlp_all.bib', 'nlp/text_normalization/tn_itn_all.bib', 'tools/tools_all.bib', 'tts/tts_all.bib', 'text_processing/text_processing_all.bib', 'core/adapters/adapter_bib.bib', 'multimodal/mm_all.bib', 'vision/vision_all.bib', ] intersphinx_mapping = { 'pytorch': ('https://pytorch.org/docs/stable', None), 'pytorch-lightning': ('https://pytorch-lightning.readthedocs.io/en/latest/', None), } # Set default flags for all classes. autodoc_default_options = {'members': None, 'undoc-members': None, 'show-inheritance': True} locale_dirs = ['locale/'] # path is example but recommended. gettext_compact = False # optional. # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # # source_suffix = ['.rst', '.md'] source_suffix = ".rst" # The master toctree document. master_doc = "index" # General information about the project. project = "NVIDIA NeMo" copyright = "© 2021-2023 NVIDIA Corporation & Affiliates. All rights reserved." author = "NVIDIA CORPORATION" # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # The short X.Y version. # version = "0.10.0" version = __version__ # The full version, including alpha/beta/rc tags. # release = "0.9.0" release = __version__ # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = None # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This patterns also effect to html_static_path and html_extra_path exclude_patterns = [] # The name of the Pygments (syntax highlighting) style to use. pygments_style = "default" ### Previous NeMo theme # # NVIDIA theme settings. # html_theme = 'nvidia_theme' # html_theme_path = ["."] # html_theme_options = { # 'display_version': True, # 'project_version': version, # 'project_name': project, # 'logo_path': None, # 'logo_only': True, # } # html_title = 'Introduction' # html_logo = html_theme_options["logo_path"] # html_sidebars = { # "**": ["navbar-logo.html", "search-field.html", "sbt-sidebar-nav.html"] # } # -- Options for HTMLHelp output ------------------------------------------ # Output file base name for HTML help builder. htmlhelp_basename = "nemodoc" ### from TLT conf.py # -- Options for HTML output ------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # # html_theme_path = [sphinx_rtd_theme.get_html_theme_path()] html_theme = "sphinx_book_theme" html_logo = os.path.join('nv_logo.png') html_title = 'NVIDIA NeMo' html_theme_options = { 'logo_only': False, 'display_version': True, # 'prev_next_buttons_location': 'bottom', # 'style_external_links': False, # 'style_nav_header_background': '#000000', # Toc options 'collapse_navigation': False, # 'sticky_navigation': False, 'navigation_depth': 10, # 'includehidden': False, # 'titles_only': False, # Sphinx Book theme, 'repository_url': 'https://github.com/NVIDIA/NeMo', 'use_repository_button': True, 'show_navbar_depth': 1, 'show_toc_level': 10, } # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_favicon = 'favicon.ico' html_static_path = ['_static'] html_last_updated_fmt = '' def setup(app): app.add_css_file('css/custom.css') app.add_js_file('js/pk_scripts.js') # html_css_files = [ # './custom.css', # ] # html_js_files = [ # './pk_scripts.js', # ] # OpenGraph settings ogp_site_url = 'https://nvidia.github.io/NeMo/' ogp_image = 'https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/main/_static/nv_logo.png' # MathJax CDN # follow recommendation here https://www.sphinx-doc.org/en/master/usage/extensions/math.html#module-sphinx.ext.mathjax mathjax_path = "https://cdn.jsdelivr.net/npm/mathjax@2/MathJax.js?config=TeX-AMS-MML_HTMLorMML" [rtd-command-info] start-time: 2024-12-19T21:09:59.750045Z, end-time: 2024-12-19T21:11:36.205015Z, duration: 96, exit-code: 0 python -m sphinx -T -b html -d _build/doctrees -D language=en . $READTHEDOCS_OUTPUT/html Running Sphinx v8.1.3 Adding req : `fiddle` to autodoc mock requirements (lib fiddle) Adding req : `huggingface_hub` to autodoc mock requirements (lib huggingface_hub>=0.24) Adding req : `numba` to autodoc mock requirements (lib numba) Skipping req : `numpy` (lib numpy>=1.22) Adding req : `onnx` to autodoc mock requirements (lib onnx>=1.7.0) Adding req : `python_dateutil` to autodoc mock requirements (lib python-dateutil) Adding req : `ruamel` to autodoc mock requirements (lib ruamel.yaml) Adding req : `scikit_learn` to autodoc mock requirements (lib scikit-learn) Adding req : `setuptools` to autodoc mock requirements (lib setuptools>=65.5.1) Adding req : `tensorboard` to autodoc mock requirements (lib tensorboard) Adding req : `text_unidecode` to autodoc mock requirements (lib text-unidecode) `torch` already added to autodoc mock requirements (lib torch) Adding req : `tqdm` to autodoc mock requirements (lib tqdm>=4.41.0) Adding req : `wget` to autodoc mock requirements (lib wget) Skipping req : `wrapt` (lib wrapt) Adding req : `braceexpand` to autodoc mock requirements (lib braceexpand) Adding req : `editdistance` to autodoc mock requirements (lib editdistance) Adding req : `einops` to autodoc mock requirements (lib einops) Adding req : `g2p_en` to autodoc mock requirements (lib g2p_en) Adding req : `jiwer` to autodoc mock requirements (lib jiwer) Adding req : `kaldi_python_io` to autodoc mock requirements (lib kaldi-python-io) Adding req : `kaldiio` to autodoc mock requirements (lib kaldiio) Adding req : `lhotse` to autodoc mock requirements (lib lhotse>=1.26.0) Adding req : `librosa` to autodoc mock requirements (lib librosa>=0.10.0) Adding req : `marshmallow` to autodoc mock requirements (lib marshmallow) Adding req : `packaging` to autodoc mock requirements (lib packaging) Adding req : `pyannote` to autodoc mock requirements (lib pyannote.core) `pyannote` already added to autodoc mock requirements (lib pyannote.metrics) Adding req : `pydub` to autodoc mock requirements (lib pydub) Adding req : `pyloudnorm` to autodoc mock requirements (lib pyloudnorm) Adding req : `resampy` to autodoc mock requirements (lib resampy) `ruamel` already added to autodoc mock requirements (lib ruamel.yaml) Adding req : `scipy` to autodoc mock requirements (lib scipy>=0.14) Adding req : `soundfile` to autodoc mock requirements (lib soundfile) Adding req : `sox` to autodoc mock requirements (lib sox) Adding req : `texterrors` to autodoc mock requirements (lib texterrors) `einops` already added to autodoc mock requirements (lib einops) `lhotse` already added to autodoc mock requirements (lib lhotse>=1.22.0) `librosa` already added to autodoc mock requirements (lib librosa>=0.10.0) Adding req : `matplotlib` to autodoc mock requirements (lib matplotlib) Adding req : `pesq` to autodoc mock requirements (lib pesq) Adding req : `pystoi` to autodoc mock requirements (lib pystoi) `scipy` already added to autodoc mock requirements (lib scipy>=0.14) `soundfile` already added to autodoc mock requirements (lib soundfile) `sox` already added to autodoc mock requirements (lib sox) Adding req : `datasets` to autodoc mock requirements (lib datasets) Adding req : `inflect` to autodoc mock requirements (lib inflect) Adding req : `pandas` to autodoc mock requirements (lib pandas) Adding req : `sacremoses` to autodoc mock requirements (lib sacremoses>=0.0.43) Adding req : `sentencepiece` to autodoc mock requirements (lib sentencepiece<1.0.0) Adding req : `fastapi` to autodoc mock requirements (lib fastapi) Adding req : `nvidia_pytriton` to autodoc mock requirements (lib nvidia-pytriton) Adding req : `pydantic_settings` to autodoc mock requirements (lib pydantic-settings) Adding req : `tensorstore` to autodoc mock requirements (lib tensorstore==0.1.45) Adding req : `uvicorn` to autodoc mock requirements (lib uvicorn) Adding req : `zarr` to autodoc mock requirements (lib zarr) Adding req : `cloudpickle` to autodoc mock requirements (lib cloudpickle) `fiddle` already added to autodoc mock requirements (lib fiddle) Adding req : `hydra_core` to autodoc mock requirements (lib hydra-core>1.3,<=1.3.2) Adding req : `omegaconf` to autodoc mock requirements (lib omegaconf<=2.3) Adding req : `pytorch_lightning` to autodoc mock requirements (lib pytorch-lightning>2.2.1) `torchmetrics` already added to autodoc mock requirements (lib torchmetrics>=0.11.0) Adding req : `transformers` to autodoc mock requirements (lib transformers) Adding req : `wandb` to autodoc mock requirements (lib wandb) Adding req : `webdataset` to autodoc mock requirements (lib webdataset>=0.2.86) Adding req : `addict` to autodoc mock requirements (lib addict) Adding req : `clip` to autodoc mock requirements (lib clip) Adding req : `decord` to autodoc mock requirements (lib decord) Adding req : `diffusers` to autodoc mock requirements (lib diffusers>=0.19.3) Adding req : `einops_exts` to autodoc mock requirements (lib einops_exts) Adding req : `imageio` to autodoc mock requirements (lib imageio) Adding req : `kornia` to autodoc mock requirements (lib kornia) Adding req : `nerfacc` to autodoc mock requirements (lib nerfacc>=0.5.3) Adding req : `open_clip_torch` to autodoc mock requirements (lib open_clip_torch==2.24.0) Adding req : `PyMCubes` to autodoc mock requirements (lib PyMCubes) Adding req : `taming_transformers` to autodoc mock requirements (lib taming-transformers) Adding req : `torchdiffeq` to autodoc mock requirements (lib torchdiffeq) Adding req : `torchsde` to autodoc mock requirements (lib torchsde) Adding req : `trimesh` to autodoc mock requirements (lib trimesh) Adding req : `accelerated_scan` to autodoc mock requirements (lib accelerated-scan) `boto3` already added to autodoc mock requirements (lib boto3) `einops` already added to autodoc mock requirements (lib einops) Adding req : `faiss_cpu` to autodoc mock requirements (lib faiss-cpu) Adding req : `fasttext` to autodoc mock requirements (lib fasttext) Adding req : `flask_restful` to autodoc mock requirements (lib flask_restful) Adding req : `ftfy` to autodoc mock requirements (lib ftfy) Adding req : `gdown` to autodoc mock requirements (lib gdown) Adding req : `h5py` to autodoc mock requirements (lib h5py) Adding req : `ijson` to autodoc mock requirements (lib ijson) Adding req : `jieba` to autodoc mock requirements (lib jieba) Adding req : `markdown2` to autodoc mock requirements (lib markdown2) `matplotlib` already added to autodoc mock requirements (lib matplotlib>=3.3.2) Adding req : `` to autodoc mock requirements (lib #megatron_core>0.6.0 # add back once mcore on pypi is compatible again) Adding req : `nltk` to autodoc mock requirements (lib nltk>=3.6.5) Adding req : `opencc` to autodoc mock requirements (lib opencc<1.1.7) Adding req : `pangu` to autodoc mock requirements (lib pangu) Adding req : `rapidfuzz` to autodoc mock requirements (lib rapidfuzz) Adding req : `rouge_score` to autodoc mock requirements (lib rouge_score) Adding req : `sacrebleu` to autodoc mock requirements (lib sacrebleu # manually install sacrebleu[ja] for Japanese support; MeCab is unsupported in Python 3.11+) Adding req : `sentence_transformers` to autodoc mock requirements (lib sentence_transformers) `tensorstore` already added to autodoc mock requirements (lib tensorstore<0.1.46) Adding req : `tiktoken` to autodoc mock requirements (lib tiktoken==0.7.0) `zarr` already added to autodoc mock requirements (lib zarr) `jiwer` already added to autodoc mock requirements (lib jiwer>=2.0.0) Adding req : `progress` to autodoc mock requirements (lib progress>=1.5) Adding req : `tabulate` to autodoc mock requirements (lib tabulate>=0.8.7) Adding req : `textdistance` to autodoc mock requirements (lib textdistance>=4.1.5) `tqdm` already added to autodoc mock requirements (lib tqdm) Adding req : `black` to autodoc mock requirements (lib black~=24.3) Adding req : `click` to autodoc mock requirements (lib click>=8.1) Adding req : `isort` to autodoc mock requirements (lib isort>5.1.0,<6.0.0) Adding req : `parameterized` to autodoc mock requirements (lib parameterized) Adding req : `pytest` to autodoc mock requirements (lib pytest) Adding req : `pytest_mock` to autodoc mock requirements (lib pytest-mock) WARNING: Invalid configuration value found: 'language = None'. Update your configuration to a valid language code. Falling back to 'en' (English). Adding req : `pytest_runner` to autodoc mock requirements (lib pytest-runner) `ruamel` already added to autodoc mock requirements (lib ruamel.yaml) Adding req : `sphinx` to autodoc mock requirements (lib sphinx) Adding req : `sphinxcontrib_bibtex` to autodoc mock requirements (lib sphinxcontrib-bibtex) `wandb` already added to autodoc mock requirements (lib wandb) `wget` already added to autodoc mock requirements (lib wget) Skipping req : `wrapt` (lib wrapt) Adding req : `attrdict` to autodoc mock requirements (lib attrdict) `einops` already added to autodoc mock requirements (lib einops) Adding req : `janome` to autodoc mock requirements (lib janome) `jieba` already added to autodoc mock requirements (lib jieba) `kornia` already added to autodoc mock requirements (lib kornia) `librosa` already added to autodoc mock requirements (lib librosa) `matplotlib` already added to autodoc mock requirements (lib matplotlib) `` already added to autodoc mock requirements (lib # pynini does not currently support aarch, disable nemo_text_processing for now) Adding req : `nemo_text_processing` to autodoc mock requirements (lib nemo_text_processing; 'arm' not in platform_machine and 'aarch' not in platform_machine) `nltk` already added to autodoc mock requirements (lib nltk) `pandas` already added to autodoc mock requirements (lib pandas) Adding req : `pypinyin` to autodoc mock requirements (lib pypinyin) Adding req : `pypinyin_dict` to autodoc mock requirements (lib pypinyin-dict) Adding req : `vllm` to autodoc mock requirements (lib vllm==0.5.3.post1) loading translations [en]... done matplotlib is not installed, social cards will not be generated making output directory... done Converting `source_suffix = '.rst'` to `source_suffix = {'.rst': 'restructuredtext'}`. checking bibtex cache... out of date parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/asr_all.bib... parsed 113 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/nlp_all.bib... parsed 35 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/text_normalization/tn_itn_all.bib... parsed 9 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tools/tools_all.bib... parsed 1 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tts/tts_all.bib... parsed 17 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/text_processing/text_processing_all.bib... WARNING: could not open bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/text_processing/text_processing_all.bib. [bibtex.bibfile_error] parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/core/adapters/adapter_bib.bib... parsed 2 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/mm_all.bib... parsed 23 entries parsing bibtex file /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/vision/vision_all.bib... parsed 2 entries loading intersphinx inventory 'pytorch' from https://pytorch.org/docs/stable/objects.inv ... loading intersphinx inventory 'pytorch-lightning' from https://pytorch-lightning.readthedocs.io/en/latest/objects.inv ... intersphinx inventory has moved: https://pytorch-lightning.readthedocs.io/en/latest/objects.inv -> https://lightning.ai/docs/pytorch/latest/objects.inv building [mo]: targets for 0 po files that are out of date writing output... building [html]: targets for 173 source files that are out of date updating environment: [new config] 173 added, 0 changed, 0 removed reading sources... [ 1%] apis reading sources... [ 1%] asr/api reading sources... [ 2%] asr/asr_language_modeling_and_customization reading sources... [ 2%] asr/configs reading sources... [ 3%] asr/datasets reading sources... [ 3%] asr/examples/kinyarwanda_asr reading sources... [ 4%] asr/intro reading sources... [ 5%] asr/models reading sources... [ 5%] asr/results reading sources... [ 6%] asr/scores reading sources... [ 6%] asr/speaker_diarization/api reading sources... [ 7%] asr/speaker_diarization/configs reading sources... [ 8%] asr/speaker_diarization/datasets reading sources... [ 8%] asr/speaker_diarization/intro reading sources... [ 9%] asr/speaker_diarization/models reading sources... [ 9%] asr/speaker_diarization/resources reading sources... [ 10%] asr/speaker_diarization/results reading sources... [ 10%] asr/speaker_recognition/api reading sources... [ 11%] asr/speaker_recognition/configs reading sources... [ 12%] asr/speaker_recognition/datasets reading sources... [ 12%] asr/speaker_recognition/intro reading sources... [ 13%] asr/speaker_recognition/models reading sources... [ 13%] asr/speaker_recognition/resources reading sources... [ 14%] asr/speaker_recognition/results reading sources... [ 14%] asr/speech_classification/configs reading sources... [ 15%] asr/speech_classification/datasets reading sources... [ 16%] asr/speech_classification/intro reading sources... [ 16%] asr/speech_classification/models reading sources... [ 17%] asr/speech_classification/resources reading sources... [ 17%] asr/speech_classification/results reading sources... [ 18%] asr/speech_intent_slot/api reading sources... [ 18%] asr/speech_intent_slot/configs reading sources... [ 19%] asr/speech_intent_slot/datasets reading sources... [ 20%] asr/speech_intent_slot/intro reading sources... [ 20%] asr/speech_intent_slot/models reading sources... [ 21%] asr/speech_intent_slot/resources reading sources... [ 21%] asr/speech_intent_slot/results reading sources... [ 22%] asr/ssl/api reading sources... [ 23%] asr/ssl/configs reading sources... [ 23%] asr/ssl/datasets reading sources... [ 24%] asr/ssl/intro reading sources... [ 24%] asr/ssl/models reading sources... [ 25%] asr/ssl/resources reading sources... [ 25%] asr/ssl/results reading sources... [ 26%] checkpoints/convert_mlm reading sources... [ 27%] checkpoints/dev_guide reading sources... [ 27%] checkpoints/dist_ckpt reading sources... [ 28%] checkpoints/intro reading sources... [ 28%] checkpoints/user_guide reading sources... [ 29%] collections reading sources... [ 29%] common/callbacks reading sources... [ 30%] common/data reading sources... [ 31%] common/intro reading sources... [ 31%] common/losses reading sources... [ 32%] common/metrics reading sources... [ 32%] common/s3_checkpointing reading sources... [ 33%] common/tokenizers reading sources... [ 34%] core/adapters/api reading sources... [ 34%] core/adapters/components reading sources... [ 35%] core/adapters/intro reading sources... [ 35%] core/api reading sources... [ 36%] core/core reading sources... [ 36%] core/exp_manager reading sources... [ 37%] core/export reading sources... [ 38%] core/neural_modules reading sources... [ 38%] core/neural_types reading sources... [ 39%] features/mixed_precision reading sources... [ 39%] features/moe reading sources... [ 40%] features/optimizations/activation_recomputation reading sources... [ 40%] features/optimizations/attention_optimizations reading sources... [ 41%] features/optimizations/communication_overlap reading sources... [ 42%] features/optimizations/cpu_offloading reading sources... [ 42%] features/optimizations/index reading sources... [ 43%] features/optimizations/sequence_packing reading sources... [ 43%] features/parallelisms reading sources... [ 44%] index reading sources... [ 45%] multimodal/api [NeMo I 2024-12-19 21:10:10 megatron_utils:39] Torch home not found, caching megatron in cwd reading sources... [ 45%] multimodal/mllm/configs reading sources... [ 46%] multimodal/mllm/datasets reading sources... [ 46%] multimodal/mllm/intro reading sources... [ 47%] multimodal/mllm/neva reading sources... [ 47%] multimodal/mllm/sequence_packing reading sources... [ 48%] multimodal/mllm/video_neva reading sources... [ 49%] multimodal/nerf/configs reading sources... [ 49%] multimodal/nerf/datasets reading sources... [ 50%] multimodal/nerf/dreamfusion reading sources... [ 50%] multimodal/nerf/intro reading sources... [ 51%] multimodal/speech_llm/api reading sources... [ 51%] multimodal/speech_llm/configs reading sources... [ 52%] multimodal/speech_llm/datasets reading sources... [ 53%] multimodal/speech_llm/intro reading sources... [ 53%] multimodal/text2img/checkpoint reading sources... [ 54%] multimodal/text2img/configs reading sources... [ 54%] multimodal/text2img/controlnet reading sources... [ 55%] multimodal/text2img/datasets reading sources... [ 55%] multimodal/text2img/dreambooth reading sources... [ 56%] multimodal/text2img/imagen reading sources... [ 57%] multimodal/text2img/insp2p reading sources... [ 57%] multimodal/text2img/intro reading sources... [ 58%] multimodal/text2img/sd reading sources... [ 58%] multimodal/text2img/sdxl_quantization reading sources... [ 59%] multimodal/vlm/checkpoint reading sources... [ 60%] multimodal/vlm/clip reading sources... [ 60%] multimodal/vlm/configs reading sources... [ 61%] multimodal/vlm/datasets reading sources... [ 61%] multimodal/vlm/intro reading sources... [ 62%] nlp/api reading sources... [ 62%] nlp/bert_pretraining reading sources... [ 63%] nlp/dialogue reading sources... [ 64%] nlp/distillation reading sources... [ 64%] nlp/entity_linking reading sources... [ 65%] nlp/glue_benchmark reading sources... [ 65%] nlp/information_retrieval reading sources... [ 66%] nlp/joint_intent_slot reading sources... [ 66%] nlp/language_modeling reading sources... [ 67%] nlp/machine_translation/machine_translation reading sources... [ 68%] nlp/megatron reading sources... [ 68%] nlp/megatron_onnx_export reading sources... [ 69%] nlp/models reading sources... [ 69%] nlp/nemo_megatron/batching reading sources... [ 70%] nlp/nemo_megatron/gpt/gpt_training reading sources... [ 71%] nlp/nemo_megatron/hiddens/hiddens_module reading sources... [ 71%] nlp/nemo_megatron/intro reading sources... [ 72%] nlp/nemo_megatron/mcore_customization reading sources... [ 72%] nlp/nemo_megatron/peft/landing_page reading sources... [ 73%] nlp/nemo_megatron/peft/quick_start reading sources... [ 73%] nlp/nemo_megatron/peft/supported_methods reading sources... [ 74%] nlp/nemo_megatron/positional_embeddings reading sources... [ 75%] nlp/nemo_megatron/retro/retro_model reading sources... [ 75%] nlp/nemo_megatron/retro_legacy/retro_model_legacy reading sources... [ 76%] nlp/nlp_model reading sources... [ 76%] nlp/punctuation_and_capitalization reading sources... [ 77%] nlp/punctuation_and_capitalization_lexical_audio reading sources... [ 77%] nlp/punctuation_and_capitalization_models reading sources... [ 78%] nlp/quantization reading sources... [ 79%] nlp/question_answering reading sources... [ 79%] nlp/spellchecking_asr_customization reading sources... [ 80%] nlp/text_classification reading sources... [ 80%] nlp/text_normalization/intro reading sources... [ 81%] nlp/text_normalization/neural_models reading sources... [ 82%] nlp/text_normalization/nn_text_normalization reading sources... [ 82%] nlp/text_normalization/text_normalization_as_tagging reading sources... [ 83%] nlp/text_normalization/wfst/intro reading sources... [ 83%] nlp/text_normalization/wfst/wfst_customization reading sources... [ 84%] nlp/text_normalization/wfst/wfst_resources reading sources... [ 84%] nlp/text_normalization/wfst/wfst_text_normalization reading sources... [ 85%] nlp/text_normalization/wfst/wfst_text_processing_deployment reading sources... [ 86%] nlp/token_classification reading sources... [ 86%] starthere/best-practices reading sources... [ 87%] starthere/fundamentals reading sources... [ 87%] starthere/intro reading sources... [ 88%] starthere/migration-guide reading sources... [ 88%] starthere/tutorials reading sources... [ 89%] tools/asr_evaluator reading sources... [ 90%] tools/comparison_tool reading sources... [ 90%] tools/ctc_segmentation reading sources... [ 91%] tools/intro reading sources... [ 91%] tools/nemo_forced_aligner reading sources... [ 92%] tools/speech_data_explorer reading sources... [ 92%] tools/speech_data_processor reading sources... [ 93%] tts/api reading sources... [ 94%] tts/checkpoints reading sources... [ 94%] tts/configs reading sources... [ 95%] tts/datasets reading sources... [ 95%] tts/g2p reading sources... [ 96%] tts/intro reading sources... [ 97%] tts/models reading sources... [ 97%] tts/resources reading sources... [ 98%] vision/checkpoint reading sources... [ 98%] vision/configs reading sources... [ 99%] vision/datasets reading sources... [ 99%] vision/intro reading sources... [100%] vision/vit WARNING: autodoc: failed to import class 'asr.models.EncDecCTCModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecCTCModelBPE' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecRNNTModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecRNNTBPEModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecClassificationModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecSpeakerLabelModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.hybrid_asr_tts_models.ASRWithTTSModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.confidence_ensemble.ConfidenceEnsembleModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.ConvASREncoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.ConvASRDecoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.ConvASRDecoderClassification' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.SpeakerDecoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.ConformerEncoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.SqueezeformerEncoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.RNNEncoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.RNNTDecoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.StatelessTransducerDecoder' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.RNNTJoint' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.SampledRNNTJoint' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.mixins.ASRBPEMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.mixins.ASRModuleMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.transcription.TranscriptionMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.transcription.TranscribeConfig' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.AudioToMelSpectrogramPreprocessor' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.AudioToMFCCPreprocessor' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.SpectrogramAugmentation' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.modules.CropOrPadSpectrogramAugmentation' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'rnnt_decoding.RNNTDecoding' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_decoding.py", line 25, in from nemo.collections.asr.parts.submodules import rnnt_beam_decoding, rnnt_greedy_decoding File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_beam_decoding.py", line 37, in from nemo.collections.asr.modules import rnnt_abstract File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'rnnt_decoding.RNNTBPEDecoding' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_decoding.py", line 25, in from nemo.collections.asr.parts.submodules import rnnt_beam_decoding, rnnt_greedy_decoding File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_beam_decoding.py", line 37, in from nemo.collections.asr.modules import rnnt_abstract File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'rnnt_greedy_decoding.GreedyRNNTInfer' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_greedy_decoding.py", line 36, in from nemo.collections.asr.modules import rnnt_abstract File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'rnnt_greedy_decoding.GreedyBatchedRNNTInfer' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_greedy_decoding.py", line 36, in from nemo.collections.asr.modules import rnnt_abstract File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'rnnt_beam_decoding.BeamRNNTInfer' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/rnnt_beam_decoding.py", line 37, in from nemo.collections.asr.modules import rnnt_abstract File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'adapters.multi_head_attention_adapter_module.MultiHeadAttentionAdapter' from module 'nemo.collections.asr.parts.submodules'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/adapters/__init__.py", line 29, in from nemo.collections.asr.parts.submodules.adapters.transformer_multi_head_attention_adapter_module import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/submodules/adapters/transformer_multi_head_attention_adapter_module.py", line 22, in from nemo.collections.asr.modules.transformer import transformer_modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/api.rst:298: ERROR: Document or section may not begin with a transition. [docutils] WARNING: autodoc: failed to import class 'asr.models.ClusteringDiarizer' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.EncDecDiarLabelModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.mixins.DiarizationMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.label_models.EncDecSpeakerLabelModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.SLUIntentSlotBPEModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.ASRModuleMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.ASRBPEMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'asr.models.SpeechEncDecSelfSupervisedModel' from module 'nemo.collections'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'mixins.mixins.ASRModuleMixin' from module 'nemo.collections.asr.parts'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/__init__.py", line 17, in from nemo.collections.asr.parts.mixins.mixins import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/parts/mixins/mixins.py", line 25, in import nemo.collections.asr.models as asr_models File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/checkpoints/intro.rst:46: WARNING: Title underline too short. Community Checkpoint Converter ----------------------------- [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/checkpoints/intro.rst:46: WARNING: Title underline too short. Community Checkpoint Converter ----------------------------- [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/collections.rst:18: WARNING: toctree contains reference to nonexisting document 'mumtimoda/speech_llm/intro' [toc.not_readable] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/core/core.rst:754: ERROR: Unexpected indentation. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/core/exp_manager.rst:325: WARNING: Explicit markup ends without a blank line; unexpected unindent. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/mixed_precision.rst:27: ERROR: "list-table" widths do not match the number of columns in table (2). .. list-table:: FP8 arguments :widths: 25 25 50 :header-rows: 1 * - Argument - Description * - transformer_engine - TE and related functionality can be enabled by setting this boolean argument to True. If this argument is not set to True, all subsequent arguments will be ignored. * - fp8 - Enables FP8 training. For transformer networks, the QKV, projection, FC1, and FC2 matrix multiplications are executed using the 4th generation H100 tensor cores with FP8 support. * - fp8_e4m3 - Training recipe format for FP8. Activations, weights, and gradient tensors use the E4M3 format. * - fp8_hybrid - Training recipe format for FP8. Activations and weight tensors use the E4M3 format, whereas gradient use the E5M2 format to satisfy the additional dynamic range requirement for backward tensors. This is the default setting. * - fp8_margin - The scaling factor for FP8 tensors can be shifted by a factor of $2 ^ {margin}$ using this argument. * - fp8_amax_history_len - Window size for amax history. The window size determines how many instances of the most recent absolute max values (amaxes) are stored per tensor. * - fp8_amax_compute_algo - The choice between “max” and “most_recent” specifies how to select an amax value from the given history. * - reduce_amax - Indicates whether or not to perform an allreduce on the amax (absolute max) values for the FP8 tensors. Since the amax is directly used to compute the scaling factor for FP8 tensors, setting this argument ensures that the scaling factors for a tensor remain synchronized across devices in multi-GPU training configurations. * - fp8_params - Indicates whether or not to store module level parameters in FP8. Enabling this option can lead to reduced memory consumption. It eliminates the need to store a copy of weights in higher precision (> half) for cases where these weights are externally maintained, such as master parameters in the optimizer. For more information, refer to the `fp8_model_init `_ API in TE. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/moe.rst:63: ERROR: Unexpected indentation. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/parallelisms.rst:4: WARNING: Duplicate explicit target name: "nemo megatron gpt config". [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/parallelisms.rst:4: WARNING: Duplicate explicit target name: "nemo megatron gpt config". [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/parallelisms.rst:4: WARNING: Duplicate explicit target name: "nemo megatron gpt config". [docutils] WARNING: autodoc: failed to import class 'ddpm.MegatronLatentDiffusion' from module 'nemo.collections.multimodal.models.text_to_image.stable_diffusion.ldm'; the following exception was raised: No module named 'open_clip' [autodoc.import_object] WARNING: autodoc: failed to import class 'dreambooth.MegatronDreamBooth' from module 'nemo.collections.multimodal.models.text_to_image.dreambooth'; the following exception was raised: No module named 'open_clip' [autodoc.import_object] WARNING: autodoc: failed to import class 'controlnet.MegatronControlNet' from module 'nemo.collections.multimodal.models.text_to_image.controlnet'; the following exception was raised: No module named 'open_clip' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/models/text_to_image/imagen/imagen.py:docstring of nemo.collections.multimodal.models.text_to_image.imagen.imagen.MegatronImagen.setup:4: WARNING: Definition list ends without a blank line; unexpected unindent. [docutils] WARNING: autodoc: failed to import class 'modules.FrozenMegatronCLIPEmbedder' from module 'nemo.collections.multimodal.modules.stable_diffusion.encoders'; the following exception was raised: No module named 'open_clip' [autodoc.import_object] WARNING: autodoc: failed to import class 'controlnet.ControlledUnetModel' from module 'nemo.collections.multimodal.models.text_to_image.controlnet'; the following exception was raised: No module named 'open_clip' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/mllm/datasets.rst:2: WARNING: Duplicate explicit target name: "huggingface". [docutils] WARNING: autodoc: failed to import class 'speech_llm.models.modular_models.ModularAudioGPTModel' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.models.modular_models.CrossAttendModularAudioGPTModel' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.models.modular_t5_models.ModularizedAudioT5Model' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.models.modular_t5_models.DecoderTextPromptModularizedAudioT5Model' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.modules.perception_modules.AudioPerceptionModule' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.modules.perception_modules.MultiAudioPerceptionModule' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.modules.TransformerCrossAttention' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.audio_text_dataset.AudioTextDataset' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.audio_text_dataset.TarredAudioTextDataset' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.audio_text_dataset.get_tarred_audio_text_dataset_from_config' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.audio_text_dataset.get_audio_text_dataset_from_config' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.lhotse_dataset.LhotseAudioQuestionAnswerDataset' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.build_dataset.build_speechllm_dataset' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] WARNING: autodoc: failed to import class 'speech_llm.data.build_dataset.build_speechllm_dataloader' from module 'nemo.collections.multimodal'; the following exception was raised: Traceback (most recent call last): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/envs/11450/lib/python3.10/site-packages/sphinx/ext/autodoc/importer.py", line 143, in import_module return importlib.import_module(modname) File "/home/docs/.asdf/installs/python/3.10.15/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 1050, in _gcd_import File "", line 1027, in _find_and_load File "", line 1006, in _find_and_load_unlocked File "", line 688, in _load_unlocked File "", line 883, in exec_module File "", line 241, in _call_with_frames_removed File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm import models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/__init__.py", line 15, in from nemo.collections.multimodal.speech_llm.models.modular_models import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/multimodal/speech_llm/models/modular_models.py", line 32, in from nemo.collections.asr.models import ASRModel, EncDecSpeakerLabelModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/__init__.py", line 15, in from nemo.collections.asr import data, losses, models, modules File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/__init__.py", line 15, in from nemo.collections.asr.models.aed_multitask_models import EncDecMultiTaskModel File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/models/aed_multitask_models.py", line 28, in from nemo.collections.asr.data.audio_to_text_lhotse_prompted import ( File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 45, in class PromptedAudioToTextLhotseDataset(torch.utils.data.Dataset): File "/home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/nemo/collections/asr/data/audio_to_text_lhotse_prompted.py", line 103, in PromptedAudioToTextLhotseDataset def _collate_tokens(self, tokens: list[list[int] | torch.Tensor]) -> tuple[torch.Tensor, torch.Tensor]: TypeError: unsupported operand type(s) for |: 'types.GenericAlias' and 'Tensor' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/configs.rst:2: WARNING: Duplicate explicit target name: "the examples". [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/intro.rst:6: ERROR: Error in "toctree" directive: invalid option block. .. toctree:: :maxdepth: 1 datasets configs api [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/text2img/sdxl_quantization.rst:141: WARNING: Explicit markup ends without a blank line; unexpected unindent. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/text2img/sdxl_quantization.rst:150: WARNING: Explicit markup ends without a blank line; unexpected unindent. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/text2img/sdxl_quantization.rst:159: WARNING: Explicit markup ends without a blank line; unexpected unindent. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/models.rst:8: WARNING: toctree contains reference to nonexisting document 'nlp/nemo_megatron/prompt_learning' [toc.not_readable] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/nemo_megatron/intro.rst:12: WARNING: toctree contains reference to nonexisting document 'nlp/nemo_megatron/prompt_learning' [toc.not_readable] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization_lexical_audio.rst:52: WARNING: duplicate label raw_data_format_punct, other instance in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization.rst /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization_lexical_audio.rst:78: WARNING: duplicate label nemo-data-format-label, other instance in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization.rst /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization_lexical_audio.rst:134: WARNING: duplicate label regular-dataset-parameters-label, other instance in /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/punctuation_and_capitalization.rst /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/fundamentals.rst:212: WARNING: Title underline too short. Fine-Tuning ---------- [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/fundamentals.rst:212: WARNING: Title underline too short. Fine-Tuning ---------- [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/fundamentals.rst:222: WARNING: Title underline too short. Where To Go Next? ----------- [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/fundamentals.rst:222: WARNING: Title underline too short. Where To Go Next? ----------- [docutils] WARNING: autodoc: failed to import class 'tts.models.FastPitchModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.MixerTTSModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.RadTTSModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.Tacotron2Model' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.SpectrogramEnhancerModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.AlignerModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.TwoStagesModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.GriffinLimModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.HifiGanModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.UnivNetModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.WaveGlowModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.AudioCodecModel' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.base.MelToSpec' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.base.SpectrogramGenerator' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'tts.models.base.Vocoder' from module 'nemo.collections'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'dataset.MixerTTSXDataset' from module 'nemo.collections.tts.data'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'dataset.TTSDataset' from module 'nemo.collections.tts.data'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] WARNING: autodoc: failed to import class 'dataset.VocoderDataset' from module 'nemo.collections.tts.data'; the following exception was raised: No module named 'seaborn' [autodoc.import_object] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/nemo_megatron/intro.rst:28: WARNING: duplicate citation for key "nlp-megatron-shoeybi2019megatron" [bibtex.duplicate_citation] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/token_classification.rst:217: WARNING: duplicate citation for key "nlp-ner-antonova2023spellmapper" [bibtex.duplicate_citation] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/token_classification.rst:217: WARNING: duplicate citation for key "nlp-ner-devlin2018bert" [bibtex.duplicate_citation] looking for now-outdated files... none found pickling environment... done checking consistency... /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/distillation.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/megatron.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/models.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/nemo_megatron/retro_legacy/retro_model_legacy.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/best-practices.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/starthere/migration-guide.rst: WARNING: document isn't included in any toctree /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/api.rst: document is referenced in multiple toctrees: ['apis', 'asr/intro'], selecting: asr/intro <- asr/api /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/machine_translation/machine_translation.rst: document is referenced in multiple toctrees: ['collections', 'nlp/models'], selecting: nlp/models <- nlp/machine_translation/machine_translation /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/megatron_onnx_export.rst: document is referenced in multiple toctrees: ['collections', 'nlp/api'], selecting: nlp/api <- nlp/megatron_onnx_export done preparing documents... WARNING: unsupported theme option 'logo_only' given WARNING: unsupported theme option 'display_version' given done copying assets... copying static files... Writing evaluated template result to /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/_readthedocs/html/_static/basic.css Writing evaluated template result to /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/_readthedocs/html/_static/language_data.js Writing evaluated template result to /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/_readthedocs/html/_static/documentation_options.js Writing evaluated template result to /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/_readthedocs/html/_static/copybutton.js copying static files: done copying extra files... copying extra files: done copying assets: done writing output... [ 1%] apis writing output... [ 1%] asr/api writing output... [ 2%] asr/asr_language_modeling_and_customization writing output... [ 2%] asr/configs writing output... [ 3%] asr/datasets writing output... [ 3%] asr/examples/kinyarwanda_asr writing output... [ 4%] asr/intro writing output... [ 5%] asr/models writing output... [ 5%] asr/results writing output... [ 6%] asr/scores writing output... [ 6%] asr/speaker_diarization/api writing output... [ 7%] asr/speaker_diarization/configs writing output... [ 8%] asr/speaker_diarization/datasets writing output... [ 8%] asr/speaker_diarization/intro writing output... [ 9%] asr/speaker_diarization/models writing output... [ 9%] asr/speaker_diarization/resources writing output... [ 10%] asr/speaker_diarization/results writing output... [ 10%] asr/speaker_recognition/api writing output... [ 11%] asr/speaker_recognition/configs writing output... [ 12%] asr/speaker_recognition/datasets writing output... [ 12%] asr/speaker_recognition/intro writing output... [ 13%] asr/speaker_recognition/models writing output... [ 13%] asr/speaker_recognition/resources writing output... [ 14%] asr/speaker_recognition/results writing output... [ 14%] asr/speech_classification/configs writing output... [ 15%] asr/speech_classification/datasets writing output... [ 16%] asr/speech_classification/intro writing output... [ 16%] asr/speech_classification/models writing output... [ 17%] asr/speech_classification/resources writing output... [ 17%] asr/speech_classification/results writing output... [ 18%] asr/speech_intent_slot/api writing output... [ 18%] asr/speech_intent_slot/configs writing output... [ 19%] asr/speech_intent_slot/datasets writing output... [ 20%] asr/speech_intent_slot/intro writing output... [ 20%] asr/speech_intent_slot/models writing output... [ 21%] asr/speech_intent_slot/resources writing output... [ 21%] asr/speech_intent_slot/results writing output... [ 22%] asr/ssl/api writing output... [ 23%] asr/ssl/configs writing output... [ 23%] asr/ssl/datasets writing output... [ 24%] asr/ssl/intro writing output... [ 24%] asr/ssl/models writing output... [ 25%] asr/ssl/resources writing output... [ 25%] asr/ssl/results writing output... [ 26%] checkpoints/convert_mlm writing output... [ 27%] checkpoints/dev_guide writing output... [ 27%] checkpoints/dist_ckpt writing output... [ 28%] checkpoints/intro writing output... [ 28%] checkpoints/user_guide writing output... [ 29%] collections writing output... [ 29%] common/callbacks writing output... [ 30%] common/data writing output... [ 31%] common/intro writing output... [ 31%] common/losses writing output... [ 32%] common/metrics writing output... [ 32%] common/s3_checkpointing writing output... [ 33%] common/tokenizers writing output... [ 34%] core/adapters/api writing output... [ 34%] core/adapters/components writing output... [ 35%] core/adapters/intro writing output... [ 35%] core/api writing output... [ 36%] core/core writing output... [ 36%] core/exp_manager writing output... [ 37%] core/export writing output... [ 38%] core/neural_modules writing output... [ 38%] core/neural_types writing output... [ 39%] features/mixed_precision writing output... [ 39%] features/moe writing output... [ 40%] features/optimizations/activation_recomputation writing output... [ 40%] features/optimizations/attention_optimizations writing output... [ 41%] features/optimizations/communication_overlap writing output... [ 42%] features/optimizations/cpu_offloading writing output... [ 42%] features/optimizations/index writing output... [ 43%] features/optimizations/sequence_packing writing output... [ 43%] features/parallelisms writing output... [ 44%] index writing output... [ 45%] multimodal/api writing output... [ 45%] multimodal/mllm/configs writing output... [ 46%] multimodal/mllm/datasets writing output... [ 46%] multimodal/mllm/intro writing output... [ 47%] multimodal/mllm/neva writing output... [ 47%] multimodal/mllm/sequence_packing writing output... [ 48%] multimodal/mllm/video_neva writing output... [ 49%] multimodal/nerf/configs writing output... [ 49%] multimodal/nerf/datasets writing output... [ 50%] multimodal/nerf/dreamfusion writing output... [ 50%] multimodal/nerf/intro writing output... [ 51%] multimodal/speech_llm/api writing output... [ 51%] multimodal/speech_llm/configs writing output... [ 52%] multimodal/speech_llm/datasets writing output... [ 53%] multimodal/speech_llm/intro writing output... [ 53%] multimodal/text2img/checkpoint writing output... [ 54%] multimodal/text2img/configs writing output... [ 54%] multimodal/text2img/controlnet writing output... [ 55%] multimodal/text2img/datasets writing output... [ 55%] multimodal/text2img/dreambooth writing output... [ 56%] multimodal/text2img/imagen writing output... [ 57%] multimodal/text2img/insp2p writing output... [ 57%] multimodal/text2img/intro writing output... [ 58%] multimodal/text2img/sd writing output... [ 58%] multimodal/text2img/sdxl_quantization writing output... [ 59%] multimodal/vlm/checkpoint writing output... [ 60%] multimodal/vlm/clip writing output... [ 60%] multimodal/vlm/configs writing output... [ 61%] multimodal/vlm/datasets writing output... [ 61%] multimodal/vlm/intro writing output... [ 62%] nlp/api writing output... [ 62%] nlp/bert_pretraining writing output... [ 63%] nlp/dialogue writing output... [ 64%] nlp/distillation writing output... [ 64%] nlp/entity_linking writing output... [ 65%] nlp/glue_benchmark writing output... [ 65%] nlp/information_retrieval writing output... [ 66%] nlp/joint_intent_slot writing output... [ 66%] nlp/language_modeling writing output... [ 67%] nlp/machine_translation/machine_translation writing output... [ 68%] nlp/megatron writing output... [ 68%] nlp/megatron_onnx_export writing output... [ 69%] nlp/models writing output... [ 69%] nlp/nemo_megatron/batching writing output... [ 70%] nlp/nemo_megatron/gpt/gpt_training writing output... [ 71%] nlp/nemo_megatron/hiddens/hiddens_module writing output... [ 71%] nlp/nemo_megatron/intro writing output... [ 72%] nlp/nemo_megatron/mcore_customization writing output... [ 72%] nlp/nemo_megatron/peft/landing_page writing output... [ 73%] nlp/nemo_megatron/peft/quick_start writing output... [ 73%] nlp/nemo_megatron/peft/supported_methods writing output... [ 74%] nlp/nemo_megatron/positional_embeddings writing output... [ 75%] nlp/nemo_megatron/retro/retro_model writing output... [ 75%] nlp/nemo_megatron/retro_legacy/retro_model_legacy writing output... [ 76%] nlp/nlp_model writing output... [ 76%] nlp/punctuation_and_capitalization writing output... [ 77%] nlp/punctuation_and_capitalization_lexical_audio writing output... [ 77%] nlp/punctuation_and_capitalization_models writing output... [ 78%] nlp/quantization writing output... [ 79%] nlp/question_answering writing output... [ 79%] nlp/spellchecking_asr_customization writing output... [ 80%] nlp/text_classification writing output... [ 80%] nlp/text_normalization/intro writing output... [ 81%] nlp/text_normalization/neural_models writing output... [ 82%] nlp/text_normalization/nn_text_normalization writing output... [ 82%] nlp/text_normalization/text_normalization_as_tagging writing output... [ 83%] nlp/text_normalization/wfst/intro writing output... [ 83%] nlp/text_normalization/wfst/wfst_customization writing output... [ 84%] nlp/text_normalization/wfst/wfst_resources writing output... [ 84%] nlp/text_normalization/wfst/wfst_text_normalization writing output... [ 85%] nlp/text_normalization/wfst/wfst_text_processing_deployment writing output... [ 86%] nlp/token_classification writing output... [ 86%] starthere/best-practices writing output... [ 87%] starthere/fundamentals writing output... [ 87%] starthere/intro writing output... [ 88%] starthere/migration-guide writing output... [ 88%] starthere/tutorials writing output... [ 89%] tools/asr_evaluator writing output... [ 90%] tools/comparison_tool writing output... [ 90%] tools/ctc_segmentation writing output... [ 91%] tools/intro writing output... [ 91%] tools/nemo_forced_aligner writing output... [ 92%] tools/speech_data_explorer writing output... [ 92%] tools/speech_data_processor writing output... [ 93%] tts/api writing output... [ 94%] tts/checkpoints writing output... [ 94%] tts/configs writing output... [ 95%] tts/datasets writing output... [ 95%] tts/g2p writing output... [ 96%] tts/intro writing output... [ 97%] tts/models writing output... [ 97%] tts/resources writing output... [ 98%] vision/checkpoint writing output... [ 98%] vision/configs writing output... [ 99%] vision/datasets writing output... [ 99%] vision/intro writing output... [100%] vision/vit /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/asr_language_modeling_and_customization.rst:564: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/asr_language_modeling_and_customization.rst:564: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v1.22.0/asset-post-v1.22.0-ctcws_scheme_1.png": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/asr_language_modeling_and_customization.rst:571: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/asr_language_modeling_and_customization.rst:571: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v1.22.0/asset-post-v1.22.0-ctcws_scheme_2.png": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/datasets.rst:466: WARNING: undefined label: 'tarred datasets' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/asr/datasets.rst:474: WARNING: undefined label: 'bucketing datasets' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/core/core.rst:293: WARNING: undefined label: 'nemo asr configuration files' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/optimizations/activation_recomputation.rst:42: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/optimizations/activation_recomputation.rst:42: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v2.0.0rc0/asset-post-activation-recomputation-exampe-2.jpg": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/optimizations/activation_recomputation.rst:49: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/optimizations/activation_recomputation.rst:49: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v2.0.0rc0/asset-post-activation-recomputation-exampe-1.jpg": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/features/optimizations/sequence_packing.rst:141: WARNING: unknown document: '../multimodal/mllm/sequence_packing' [ref.doc] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/nerf/dreamfusion.rst:253: WARNING: Lexing literal_block '_target_: nemo.collections.multimodal.models.nerf.dreamfusion.DreamFusion\ndefaults:\n - nerf: torchngp\n - background: torchngp\n - material: basic_shading\n - renderer: nvdiffrast # (1)\n - guidance: sd_huggingface\n - optim: adan\n - loss: dmtet # (2)\n - data: data\n - _self_\n\n### model options\nresume_from_checkpoint: "/results/DreamFusion/checkpoints/DreamFusion-step\\=10000-last.ckpt" # (3)\nprompt: \'a hamburger\'\nnegative_prompt: \'\'\nfront_prompt: \', front view\'\nside_prompt: \', side view\'\nback_prompt: \', back view\'\nupdate_extra_interval: 16\nguidance_scale: 100\nexport_video: False\n\niters: ${trainer.max_steps}\nlatent_iter_ratio: 0.0\nalbedo_iter_ratio: 0\nmin_ambient_ratio: 0.1\ntextureless_ratio: 0.2\n\ndata:\n train_dataset:\n width: 512 # (4)\n height: 512 # (4)\n val_dataset:\n width: 800\n height: 800\n test_dataset:\n width: 800\n height: 800' as "yaml" resulted in an error at token: '\\'. Retrying in relaxed mode. [misc.highlighting_failure] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/intro.rst:24: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/intro.rst:24: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v1.23.0/salm.png": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/intro.rst:33: WARNING: Could not obtain image size. :scale: option is ignored. /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/speech_llm/intro.rst:33: WARNING: Cannot scale image! Could not get size from "https://github.com/NVIDIA/NeMo/releases/download/v1.23.0/bestow.png": Requires Python Imaging Library. [docutils] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/text2img/imagen.rst:78: WARNING: could not find bibtex key "mm-models-imagen-nichol2021improved" [bibtex.key_not_found] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/multimodal/text2img/insp2p.rst:7: WARNING: citation not found: InstructPix2Pix [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/nlp/nemo_megatron/gpt/gpt_training.rst:235: WARNING: undefined label: 'promptlearning' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tts/checkpoints.rst:26: WARNING: undefined label: 'ngc tts models' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tts/checkpoints.rst:37: WARNING: undefined label: 'ngc tts models' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tts/checkpoints.rst:92: WARNING: undefined label: 'nemo tts collection api' [ref.ref] /home/docs/checkouts/readthedocs.org/user_builds/nvidia-nemo/checkouts/11450/docs/source/tts/g2p.rst:134: WARNING: could not find bibtex key "g2p--devlin2018bert" [bibtex.key_not_found] generating indices... genindex done highlighting module code... [ 1%] nemo.collections.asr.data.audio_to_text highlighting module code... [ 3%] nemo.collections.asr.data.text_to_text highlighting module code... [ 4%] nemo.collections.asr.parts.mixins.interctc_mixin highlighting module code... [ 6%] nemo.collections.asr.parts.preprocessing.perturb highlighting module code... [ 7%] nemo.collections.asr.parts.submodules.adapters.multi_head_attention_adapter_module highlighting module code... [ 8%] nemo.collections.asr.parts.submodules.ctc_beam_decoding highlighting module code... [ 10%] nemo.collections.asr.parts.submodules.ctc_decoding highlighting module code... [ 11%] nemo.collections.asr.parts.submodules.ctc_greedy_decoding highlighting module code... [ 12%] nemo.collections.asr.parts.submodules.jasper highlighting module code... [ 14%] nemo.collections.asr.parts.utils.rnnt_utils highlighting module code... [ 15%] nemo.collections.common.data.dataset highlighting module code... [ 17%] nemo.collections.common.losses.aggregator highlighting module code... [ 18%] nemo.collections.common.losses.cross_entropy highlighting module code... [ 19%] nemo.collections.common.losses.mse_loss highlighting module code... [ 21%] nemo.collections.common.losses.smoothed_cross_entropy highlighting module code... [ 22%] nemo.collections.common.losses.spanning_loss highlighting module code... [ 24%] nemo.collections.common.metrics.perplexity highlighting module code... [ 25%] nemo.collections.common.parts.adapter_modules highlighting module code... [ 26%] nemo.collections.common.tokenizers.huggingface.auto_tokenizer highlighting module code... [ 28%] nemo.collections.common.tokenizers.sentencepiece_tokenizer highlighting module code... [ 29%] nemo.collections.common.tokenizers.tokenizer_spec highlighting module code... [ 31%] nemo.collections.multimodal.data.common.webdataset highlighting module code... [ 32%] nemo.collections.multimodal.data.dreambooth.dreambooth_dataset highlighting module code... [ 33%] nemo.collections.multimodal.models.text_to_image.imagen.imagen highlighting module code... [ 35%] nemo.collections.multimodal.models.text_to_image.stable_diffusion.ldm.autoencoder highlighting module code... [ 36%] nemo.collections.multimodal.modules.imagen.diffusionmodules.nets highlighting module code... [ 38%] nemo.collections.multimodal.modules.imagen.encoder.t5encoder highlighting module code... [ 39%] nemo.collections.multimodal.modules.stable_diffusion.diffusionmodules.openaimodel highlighting module code... [ 40%] nemo.collections.nlp.data.language_modeling.megatron.base_prompt_learning_dataset highlighting module code... [ 42%] nemo.collections.nlp.data.language_modeling.megatron.bert_dataset highlighting module code... [ 43%] nemo.collections.nlp.data.language_modeling.megatron.blendable_dataset highlighting module code... [ 44%] nemo.collections.nlp.data.language_modeling.megatron.gpt_dataset highlighting module code... [ 46%] nemo.collections.nlp.data.language_modeling.megatron.gpt_sft_chat_dataset highlighting module code... [ 47%] nemo.collections.nlp.data.language_modeling.megatron.gpt_sft_dataset highlighting module code... [ 49%] nemo.collections.nlp.data.language_modeling.megatron.retro_dataset highlighting module code... [ 50%] nemo.collections.nlp.data.language_modeling.megatron.t5_dataset highlighting module code... [ 51%] nemo.collections.nlp.data.language_modeling.megatron.t5_prompt_learning_dataset highlighting module code... [ 53%] nemo.collections.nlp.data.language_modeling.megatron.ul2_dataset highlighting module code... [ 54%] nemo.collections.nlp.models.language_modeling.megatron.bert.bert_model highlighting module code... [ 56%] nemo.collections.nlp.models.language_modeling.megatron.gpt_model highlighting module code... [ 57%] nemo.collections.nlp.models.language_modeling.megatron_bart_model highlighting module code... [ 58%] nemo.collections.nlp.models.language_modeling.megatron_base_model highlighting module code... [ 60%] nemo.collections.nlp.models.language_modeling.megatron_bert_model highlighting module code... [ 61%] nemo.collections.nlp.models.language_modeling.megatron_gpt_adapter_model highlighting module code... [ 62%] nemo.collections.nlp.models.language_modeling.megatron_gpt_model highlighting module code... [ 64%] nemo.collections.nlp.models.language_modeling.megatron_gpt_prompt_learning_model highlighting module code... [ 65%] nemo.collections.nlp.models.language_modeling.megatron_gpt_sft_model highlighting module code... [ 67%] nemo.collections.nlp.models.language_modeling.megatron_retrieval_model highlighting module code... [ 68%] nemo.collections.nlp.models.language_modeling.megatron_t5_adapter_model highlighting module code... [ 69%] nemo.collections.nlp.models.language_modeling.megatron_t5_model highlighting module code... [ 71%] nemo.collections.nlp.modules.common.megatron.module highlighting module code... [ 72%] nemo.collections.nlp.modules.common.megatron.retrieval_token_level_encoder_decoder highlighting module code... [ 74%] nemo.collections.nlp.modules.common.megatron.token_level_encoder_decoder highlighting module code... [ 75%] nemo.collections.nlp.parts.mixins.nlp_adapter_mixins highlighting module code... [ 76%] nemo.core.classes.common highlighting module code... [ 78%] nemo.core.classes.exportable highlighting module code... [ 79%] nemo.core.classes.mixins.access_mixins highlighting module code... [ 81%] nemo.core.classes.mixins.adapter_mixin_strategies highlighting module code... [ 82%] nemo.core.classes.mixins.hf_io_mixin highlighting module code... [ 83%] nemo.core.classes.modelPT highlighting module code... [ 85%] nemo.core.classes.module highlighting module code... [ 86%] nemo.core.connectors.save_restore_connector highlighting module code... [ 88%] nemo.core.neural_types.axes highlighting module code... [ 89%] nemo.core.neural_types.comparison highlighting module code... [ 90%] nemo.core.neural_types.elements highlighting module code... [ 92%] nemo.core.neural_types.neural_type highlighting module code... [ 93%] nemo.core.optim.lr_scheduler highlighting module code... [ 94%] nemo.core.optim.optimizers highlighting module code... [ 96%] nemo.utils.exp_manager highlighting module code... [ 97%] pytorch_lightning.LightningModule.trainer highlighting module code... [ 99%] pytorch_lightning.utilities highlighting module code... [100%] torch writing additional pages... search done copying images... [ 1%] asr/images/kinyarwanda_from_scratch.png copying images... [ 2%] asr/images/kinyarwanda_finetuning.png copying images... [ 4%] asr/images/conformer_ctc.png copying images... [ 5%] asr/images/hat.png copying images... [ 6%] asr/images/hybrid_asr_tts_model.png copying images... [ 8%] asr/images/conf-ensembles-overview.png copying images... [ 9%] asr/images/jasper_vertical.png copying images... [ 10%] asr/images/quartz_vertical.png copying images... [ 11%] asr/images/citrinet_vertical.png copying images... [ 12%] asr/images/squeezeformer.png copying images... [ 14%] asr/speaker_diarization/images/msdd_train_and_infer.png copying images... [ 15%] asr/speaker_diarization/images/asr_sd_diagram.png copying images... [ 16%] asr/speaker_diarization/images/sd_pipeline.png copying images... [ 18%] asr/speaker_diarization/images/ms_trade_off.png copying images... [ 19%] asr/speaker_diarization/images/data_flow.png copying images... [ 20%] asr/speaker_diarization/images/scale_weight_cnn.png copying images... [ 21%] asr/speaker_diarization/images/weighted_sum.png copying images... [ 22%] asr/speaker_diarization/images/sequence_model.png copying images... [ 24%] asr/speaker_recognition/images/titanet_network.png copying images... [ 25%] asr/speaker_recognition/images/ICASPP_SpeakerNet.png copying images... [ 26%] asr/speech_classification/images/matchboxnet_vertical.png copying images... [ 28%] asr/speech_classification/images/marblenet_vertical.png copying images... [ 29%] asr/speech_intent_slot/images/example.png copying images... [ 30%] asr/speech_intent_slot/images/framework.png copying images... [ 31%] core/whyntypes.gif copying images... [ 32%] nlp/nemo_megatron/images/tp_comm_overlap.png copying images... [ 34%] nlp/nemo_megatron/images/pp_comm_overlap.png copying images... [ 35%] nlp/nemo_megatron/images/ddp.gif copying images... [ 36%] nlp/nemo_megatron/images/tp.gif copying images... [ 38%] nlp/nemo_megatron/images/pp.gif copying images... [ 39%] nlp/nemo_megatron/images/ep.png copying images... [ 40%] nlp/nemo_megatron/images/sp.gif copying images... [ 41%] nlp/nemo_megatron/images/pnom.gif copying images... [ 42%] multimodal/mllm/images/llava_arch.jpg copying images... [ 44%] multimodal/nerf/images/dreamfusion_model_overview.png copying images... [ 45%] multimodal/text2img/images/controlnet-structure.png copying images... [ 46%] multimodal/text2img/images/imagen_arch.png copying images... [ 48%] multimodal/vlm/images/clip_arch.png copying images... [ 49%] nlp/dialogue_UML.png copying images... [ 50%] nlp/entity_linking_overview.jpg copying images... [ 51%] nlp/nemo_megatron/hiddens/images/hiddens-wb-logging.png copying images... [ 52%] nlp/nemo_megatron/mr1.png copying images... [ 54%] nlp/nemo_megatron/mr2.png copying images... [ 55%] nlp/nemo_megatron/customization_module.png copying images... [ 56%] nlp/nemo_megatron/customization_forward.png copying images... [ 57%] nlp/nemo_megatron/retro_legacy/images/arch.png copying images... [ 59%] nlp/question_answering_arch.png copying images... [ 60%] nlp/text_normalization/wfst/images/task_overview.png copying images... [ 61%] nlp/text_normalization/wfst/images/shallow_fusion.png copying images... [ 62%] nlp/text_normalization/wfst/images/deployment_pipeline.png copying images... [ 64%] tools/images/sde_base_stats.png copying images... [ 65%] tools/images/sde_player.png copying images... [ 66%] tools/images/scrsh_2.png copying images... [ 68%] tools/images/scrsh_3.png copying images... [ 69%] tools/images/scrsh_4.png copying images... [ 70%] tools/images/scrsh_5.png copying images... [ 71%] tools/images/scrsh_9.png copying images... [ 72%] tools/images/scr_10.png copying images... [ 74%] tools/images/scr_11.png copying images... [ 75%] tools/images/sde_words.png copying images... [ 76%] tools/images/sde_utterances.png copying images... [ 78%] tools/images/datatable_toggle.png copying images... [ 79%] tools/images/datatable_sort.png copying images... [ 80%] tools/images/datatable_filter.png copying images... [ 81%] tools/images/sde_mls_words.png copying images... [ 82%] tools/images/sde_mls_cer.png copying images... [ 84%] tools/images/sde_mls_player.png copying images... [ 85%] tts/images/data_labeling_pipeline.png copying images... [ 86%] tts/images/fastpitch_model.png copying images... [ 88%] tts/images/mixertts_model.png copying images... [ 89%] tts/images/radtts_model.png copying images... [ 90%] tts/images/tacotron2_model.png copying images... [ 91%] tts/images/hifigan_g_model.png copying images... [ 92%] tts/images/hifigan_d_model.png copying images... [ 94%] tts/images/univnet_model.png copying images... [ 95%] tts/images/waveglow_model.png copying images... [ 96%] tts/images/radaligner_model.png copying images... [ 98%] tts/images/vits_model.png copying images... [ 99%] tts/images/audiocodec_model.png copying images... [100%] vision/images/vit_arch.png dumping search index in English (code: en)... done dumping object inventory... done build succeeded, 143 warnings. The HTML pages are in ../../_readthedocs/html.