Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A error when compose LG #1318

Open
teamtee opened this issue Jan 3, 2025 · 7 comments
Open

A error when compose LG #1318

teamtee opened this issue Jan 3, 2025 · 7 comments

Comments

@teamtee
Copy link

teamtee commented Jan 3, 2025

Error description

2025-01-03 09:40:32,765 INFO [compile_hlg.py:66] Connecting LG
2025-01-03 09:40:32,765 INFO [compile_hlg.py:68] LG shape after k2.connect: (0, None)
2025-01-03 09:40:32,765 INFO [compile_hlg.py:70] <class 'torch.Tensor'>
2025-01-03 09:40:32,765 INFO [compile_hlg.py:71] Determinizing LG
2025-01-03 09:40:32,765 INFO [compile_hlg.py:74] <class 'torch.Tensor'>
2025-01-03 09:40:32,765 INFO [compile_hlg.py:76] Connecting LG after k2.determinize
2025-01-03 09:40:32,765 INFO [compile_hlg.py:79] Removing disambiguation symbols on LG
Traceback (most recent call last):
  File "./local/compile_hlg.py", line 136, in <module>
    main()
  File "./local/compile_hlg.py", line 126, in main
    HLG = compile_HLG(lang_dir)
  File "./local/compile_hlg.py", line 87, in compile_HLG
    assert isinstance(LG.aux_labels, k2.RaggedTensor)
AssertionError

Env

PyTorch version: 2.0.1+cu118
Is debug build: False
CUDA used to build PyTorch: 11.8
ROCM used to build PyTorch: N/A

OS: Rocky Linux 8.7 (Green Obsidian) (x86_64)
GCC version: (GCC) 8.5.0 20210514 (Red Hat 8.5.0-18)
Clang version: Could not collect
CMake version: version 3.25.0
Libc version: glibc-2.28

Python version: 3.8.20 (default, Oct  3 2024, 15:24:27)  [GCC 11.2.0] (64-bit runtime)
Python platform: Linux-4.18.0-425.3.1.el8.x86_64-x86_64-with-glibc2.17
Is CUDA available: True
CUDA runtime version: 12.2.140
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: 
GPU 0: NVIDIA A800-SXM4-80GB
GPU 1: NVIDIA A800-SXM4-80GB
GPU 2: NVIDIA A800-SXM4-80GB
GPU 3: NVIDIA A800-SXM4-80GB
GPU 4: NVIDIA A800-SXM4-80GB
GPU 5: NVIDIA A800-SXM4-80GB
GPU 6: NVIDIA A800-SXM4-80GB
GPU 7: NVIDIA A800-SXM4-80GB

Nvidia driver version: 535.104.05
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True

CPU:
Architecture:        x86_64
CPU op-mode(s):      32-bit, 64-bit
Byte Order:          Little Endian
CPU(s):              128
On-line CPU(s) list: 0-127
Thread(s) per core:  2
Core(s) per socket:  32
Socket(s):           2
NUMA node(s):        2
Vendor ID:           GenuineIntel
CPU family:          6
Model:               106
Model name:          Intel(R) Xeon(R) Platinum 8378A CPU @ 3.00GHz
Stepping:            6
CPU MHz:             3000.000
CPU max MHz:         3000.0000
CPU min MHz:         800.0000
BogoMIPS:            6000.00
Virtualization:      VT-x
L1d cache:           48K
L1i cache:           32K
L2 cache:            1280K
L3 cache:            49152K
NUMA node0 CPU(s):   0-31,64-95
NUMA node1 CPU(s):   32-63,96-127
Flags:               fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 invpcid_single intel_ppin ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect wbnoinvd dtherm arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid fsrm md_clear pconfig flush_l1d arch_capabilities

Versions of relevant libraries:
[pip3] k2==1.24.4.dev20241029+cuda11.8.torch2.0.1
[pip3] numpy==1.24.4
[pip3] torch==2.0.1+cu118
[pip3] torchaudio==2.0.2+cu118
[pip3] torchvision==0.15.2+cu118
[conda] k2                        1.24.4.dev20241029+cuda11.8.torch2.0.1          pypi_0    pypi
[conda] numpy                     1.24.4                   pypi_0    pypi
[conda] torch                     2.0.1+cu118              pypi_0    pypi
[conda] torchaudio                2.0.2+cu118              pypi_0    pypi
[conda] torchvision               0.15.2+cu118             pypi_0    pypi
@csukuangfj
Copy link
Collaborator

There is a 0 in the second line of the log.

Please check your L or G is not empty.

@teamtee
Copy link
Author

teamtee commented Jan 3, 2025

Thanks,I will check.I just install the k2/icefall and run the script yesno follow the tutorials.

@csukuangfj
Copy link
Collaborator

Please check the output of each stage in prepare.sh.

@teamtee
Copy link
Author

teamtee commented Jan 3, 2025

OK,the all input is here

(slam-mala) (slam-llm) [yangui.fang_sx@d8-hpc-gpu-029 ASR]$ ./prepare.sh 
2025-01-03 11:35:43 (prepare.sh:27:main) dl_dir: /hpc_stor01/home/yangui.fang_sx/workingspace/github/icefall/egs/yesno/ASR/download
2025-01-03 11:35:43 (prepare.sh:30:main) Stage 0: Download data
2025-01-03 11:35:43 (prepare.sh:39:main) Stage 1: Prepare yesno manifest
2025-01-03 11:35:47 (prepare.sh:45:main) Stage 2: Compute fbank for yesno
2025-01-03 11:35:51,406 INFO [compute_fbank_yesno.py:63] train already exists - skipping.
2025-01-03 11:35:51,408 INFO [compute_fbank_yesno.py:63] test already exists - skipping.
2025-01-03 11:35:51 (prepare.sh:51:main) Stage 3: Prepare lang
2025-01-03 11:35:56,959 INFO [prepare_lang_fst.py:174] Building standard CTC topology
2025-01-03 11:35:56,962 INFO [prepare_lang_fst.py:183] Building L
2025-01-03 11:35:56,963 INFO [prepare_lang_fst.py:191] Building HL
2025-01-03 11:35:56,977 INFO [prepare_lang_fst.py:201] Skip building HLG
2025-01-03 11:35:57 (prepare.sh:67:main) Stage 4: Prepare G
2025-01-03 11:35:57 (prepare.sh:93:main) Stage 5: Compile HLG
2025-01-03 11:35:59,494 INFO [compile_hlg.py:124] Processing data/lang_phone
2025-01-03 11:35:59,495 INFO [lexicon.py:168] Loading pre-compiled data/lang_phone/Linv.pt
2025-01-03 11:35:59,502 INFO [compile_hlg.py:48] Building ctc_topo. max_token_id: 3
2025-01-03 11:35:59,505 INFO [compile_hlg.py:52] Loading G.fst.txt
2025-01-03 11:35:59,509 INFO [compile_hlg.py:62] Intersecting L and G
2025-01-03 11:35:59,512 INFO [compile_hlg.py:64] LG shape: (0, None)
2025-01-03 11:35:59,512 INFO [compile_hlg.py:66] Connecting LG
2025-01-03 11:35:59,512 INFO [compile_hlg.py:68] LG shape after k2.connect: (0, None)
2025-01-03 11:35:59,512 INFO [compile_hlg.py:70] <class 'torch.Tensor'>
2025-01-03 11:35:59,512 INFO [compile_hlg.py:71] Determinizing LG
2025-01-03 11:35:59,512 INFO [compile_hlg.py:74] <class 'torch.Tensor'>
2025-01-03 11:35:59,512 INFO [compile_hlg.py:76] Connecting LG after k2.determinize
2025-01-03 11:35:59,512 INFO [compile_hlg.py:79] Removing disambiguation symbols on LG
Traceback (most recent call last):
  File "./local/compile_hlg.py", line 136, in <module>
    main()
  File "./local/compile_hlg.py", line 126, in main
    HLG = compile_HLG(lang_dir)
  File "./local/compile_hlg.py", line 87, in compile_HLG
    assert isinstance(LG.aux_labels, k2.RaggedTensor)
AssertionError

@teamtee
Copy link
Author

teamtee commented Jan 3, 2025

And I note the icefall/egs/yesno/ASR/data/lm/G.fst.txt is empty

@csukuangfj
Copy link
Collaborator

OK,the all input is here

Please check the input files by yourself.

You need to use ls -lh to check that.

Note that the generated files are not the logs from the terminal. You need to check the files in the disk manually.

@csukuangfj
Copy link
Collaborator

Please debug it stage by stage by yourself.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants