{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":662024634,"defaultBranch":"main","name":"kNN-TL","ownerLogin":"NLP2CT","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-07-04T07:40:30.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/57902470?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1688456441.0","currentOid":""},"activityList":{"items":[{"before":"fa39ef06357167207d7c35a84b047aaedf37f070","after":"275945104d5a08d737d8e67c4e609dfff3a37317","ref":"refs/heads/main","pushedAt":"2023-07-27T03:43:48.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"sudanl","name":"Shudong Liu","path":"/sudanl","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57032474?s=80&v=4"},"commit":{"message":"Update README.md","shortMessageHtmlLink":"Update README.md"}},{"before":"ce485e80acd366a15acbd34eb62eb4f968969295","after":"fa39ef06357167207d7c35a84b047aaedf37f070","ref":"refs/heads/main","pushedAt":"2023-07-27T03:42:00.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"sudanl","name":"Shudong Liu","path":"/sudanl","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57032474?s=80&v=4"},"commit":{"message":"\tmodified: README.md","shortMessageHtmlLink":"\tmodified: README.md"}},{"before":"913333ddb1d1ca7c41dc007d6af3d295f57f191a","after":"ce485e80acd366a15acbd34eb62eb4f968969295","ref":"refs/heads/main","pushedAt":"2023-07-04T08:56:39.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"sudanl","name":"Shudong Liu","path":"/sudanl","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57032474?s=80&v=4"},"commit":{"message":"\tmodified: .DS_Store\n\tmodified: README.md\n\tnew file: inference-scripts/build_parent_datastore.sh\n\tnew file: inference-scripts/inference_child.sh\n\tnew file: inference-scripts/inference_pseudo_parent.sh\n\tnew file: kNN-TL/.circleci/config.yml\n\tnew file: kNN-TL/.github/CODEOWNERS\n\tnew file: kNN-TL/.github/ISSUE_TEMPLATE.md\n\tnew file: kNN-TL/.github/ISSUE_TEMPLATE/bug_report.md\n\tnew file: kNN-TL/.github/ISSUE_TEMPLATE/documentation.md\n\tnew file: kNN-TL/.github/ISSUE_TEMPLATE/feature_request.md\n\tnew file: kNN-TL/.github/ISSUE_TEMPLATE/how-to-question.md\n\tnew file: kNN-TL/.github/PULL_REQUEST_TEMPLATE.md\n\tnew file: kNN-TL/.github/stale.yml\n\tnew file: kNN-TL/.github/workflows/build.yml\n\tnew file: kNN-TL/.github/workflows/release.yml\n\tnew file: kNN-TL/.gitignore\n\tnew file: kNN-TL/.gitmodules\n\tnew file: kNN-TL/.isort.cfg\n\tnew file: kNN-TL/.pre-commit-config.yaml\n\tnew file: kNN-TL/CODE_OF_CONDUCT.md\n\tnew file: kNN-TL/CONTRIBUTING.md\n\tnew file: kNN-TL/LICENSE\n\tnew file: kNN-TL/README.md\n\tnew file: kNN-TL/RELEASE.md\n\tnew file: kNN-TL/docs/Makefile\n\tnew file: kNN-TL/docs/_static/theme_overrides.css\n\tnew file: kNN-TL/docs/command_line_tools.rst\n\tnew file: kNN-TL/docs/conf.py\n\tnew file: kNN-TL/docs/criterions.rst\n\tnew file: kNN-TL/docs/data.rst\n\tnew file: kNN-TL/docs/docutils.conf\n\tnew file: kNN-TL/docs/fairseq.gif\n\tnew file: kNN-TL/docs/fairseq_logo.png\n\tnew file: kNN-TL/docs/getting_started.rst\n\tnew file: kNN-TL/docs/hydra_integration.md\n\tnew file: kNN-TL/docs/index.rst\n\tnew file: kNN-TL/docs/lr_scheduler.rst\n\tnew file: kNN-TL/docs/make.bat\n\tnew file: kNN-TL/docs/models.rst\n\tnew file: kNN-TL/docs/modules.rst\n\tnew file: kNN-TL/docs/optim.rst\n\tnew file: kNN-TL/docs/overview.rst\n\tnew file: kNN-TL/docs/requirements.txt\n\tnew file: kNN-TL/docs/tasks.rst\n\tnew file: kNN-TL/docs/tutorial_classifying_names.rst\n\tnew file: kNN-TL/docs/tutorial_simple_lstm.rst\n\tnew file: kNN-TL/examples/.gitignore\n\tnew file: kNN-TL/examples/MMPT/.gitignore\n\tnew file: kNN-TL/examples/MMPT/CONFIG.md\n\tnew file: kNN-TL/examples/MMPT/DATASET.md\n\tnew file: kNN-TL/examples/MMPT/README.md\n\tnew file: kNN-TL/examples/MMPT/endtask.md\n\tnew file: kNN-TL/examples/MMPT/locallaunch.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/datasets/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/datasets/fairseqmmdataset.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/datasets/mmdataset.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/evaluators/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/evaluators/evaluator.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/evaluators/metric.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/evaluators/predictor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/losses/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/losses/fairseqmmloss.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/losses/loss.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/losses/nce.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/models/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/models/fairseqmmmodel.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/models/mmfusion.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/models/mmfusionnlg.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/models/transformermodel.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/modules/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/modules/mm.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/modules/retri.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/modules/vectorpool.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/dedupprocessor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/dsprocessor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/how2processor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/how2retriprocessor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/models/s3dg.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/processors/processor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/fairseqmmtask.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/milncetask.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/retritask.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/task.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/tasks/vlmtask.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/utils/__init__.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/utils/load_config.py\n\tnew file: kNN-TL/examples/MMPT/mmpt/utils/shardedtensor.py\n\tnew file: kNN-TL/examples/MMPT/mmpt_cli/localjob.py\n\tnew file: kNN-TL/examples/MMPT/mmpt_cli/predict.py\n\tnew file: kNN-TL/examples/MMPT/pretraining.md\n\tnew file: kNN-TL/examples/MMPT/projects/mfmmlm.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/mmfusionmtm.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/coin.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/crosstask.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/how2.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_coin.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_crosstask.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_crosstask_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_vtt.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_vttqa.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_youcook.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/test_youcookcap.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/vtt.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/vttqa.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/youcook.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/mtm/vlm/youcookcap.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/coin_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/crosstask_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/how2.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_coin_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_coin_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_crosstask_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_crosstask_zs_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_didemo_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_vtt_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_vtt_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_vttqa_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_vttqa_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_youcook_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/test_youcook_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/vtt_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/vttqa_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoclip/youcook_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/retri/videoretri.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/coin.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/coin_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/crosstask.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/crosstask_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/default.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/ft.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/how2.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_coin.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_coin_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_coin_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_crosstask.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_crosstask_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_crosstask_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_crosstask_zs_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_didemo_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vtt.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vtt_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vtt_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vttqa.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vttqa_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_vttqa_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_youcook.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_youcook_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_youcook_zs.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/test_youcookcap.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/vtt.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/vtt_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/vttqa.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/vttqa_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/youcook.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/youcook_videoclip.yaml\n\tnew file: kNN-TL/examples/MMPT/projects/task/youcookcap.yaml\n\tnew file: kNN-TL/examples/MMPT/scripts/text_token_extractor/configs/bert-base-uncased.yaml\n\tnew file: kNN-TL/examples/MMPT/scripts/text_token_extractor/pretokenization.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/extract.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/how2/s3d.sh\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/model.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/pathbuilder.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/preprocessing.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/random_sequence_shuffler.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/shard_feature.py\n\tnew file: kNN-TL/examples/MMPT/scripts/video_feature_extractor/videoreader.py\n\tnew file: kNN-TL/examples/MMPT/setup.py\n\tnew file: kNN-TL/examples/MMPT/videoclip.png\n\tnew file: kNN-TL/examples/MMPT/vlm.png\n\tnew file: kNN-TL/examples/__init__.py\n\tnew file: kNN-TL/examples/adaptive_span/README.md\n\tnew file: kNN-TL/examples/adaptive_span/__init__.py\n\tnew file: kNN-TL/examples/adaptive_span/adagrad_with_grad_clip.py\n\tnew file: kNN-TL/examples/adaptive_span/adaptive_span_attention.py\n\tnew file: kNN-TL/examples/adaptive_span/adaptive_span_loss.py\n\tnew file: kNN-TL/examples/adaptive_span/adaptive_span_model.py\n\tnew file: kNN-TL/examples/adaptive_span/adaptive_span_model_wrapper.py\n\tnew file: kNN-TL/examples/adaptive_span/truncated_bptt_lm_task.py\n\tnew file: kNN-TL/examples/attention_head_selection/README.md\n\tnew file: kNN-TL/examples/attention_head_selection/src/__init__.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/data/__init__.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/data/speech_to_text_dataset_with_domain.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/loss/__init__.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/loss/attention_head_selection.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/models/__init__.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/models/head_selection_s2t_transformer.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/models/head_selection_transformer.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/modules/__init__.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/modules/attn_head_selector.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/modules/head_selection_transformer_layer.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/modules/multihead_attention_selection.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/modules/multihead_functional.py\n\tnew file: kNN-TL/examples/attention_head_selection/src/speech_to_text_head_selection.py\n\tnew file: kNN-TL/examples/backtranslation/README.md\n\tnew file: kNN-TL/examples/backtranslation/deduplicate_lines.py\n\tnew file: kNN-TL/examples/backtranslation/extract_bt_data.py\n\tnew file: kNN-TL/examples/backtranslation/prepare-de-monolingual.sh\n\tnew file: kNN-TL/examples/backtranslation/prepare-wmt18en2de.sh\n\tnew file: kNN-TL/examples/backtranslation/sacrebleu.sh\n\tnew file: kNN-TL/examples/backtranslation/tokenized_bleu.sh\n\tnew file: kNN-TL/examples/bart/README.glue.md\n\tnew file: kNN-TL/examples/bart/README.md\n\tnew file: kNN-TL/examples/bart/README.summarization.md\n\tnew file: kNN-TL/examples/bart/summarize.py\n\tnew file: kNN-TL/examples/byte_level_bpe/README.md\n\tnew file: kNN-TL/examples/byte_level_bpe/get_bitext.py\n\tnew file: kNN-TL/examples/byte_level_bpe/get_data.sh\n\tnew file: kNN-TL/examples/byte_level_bpe/gru_transformer.py\n\tnew file: kNN-TL/examples/camembert/README.md\n\tnew file: kNN-TL/examples/constrained_decoding/README.md\n\tnew file: kNN-TL/examples/constrained_decoding/normalize.py\n\tnew file: kNN-TL/examples/constrained_decoding/tok.py\n\tnew file: kNN-TL/examples/conv_seq2seq/README.md\n\tnew file: kNN-TL/examples/criss/README.md\n\tnew file: kNN-TL/examples/criss/download_and_preprocess_flores_test.sh\n\tnew file: kNN-TL/examples/criss/download_and_preprocess_tatoeba.sh\n\tnew file: kNN-TL/examples/criss/mining/mine.py\n\tnew file: kNN-TL/examples/criss/mining/mine_example.sh\n\tnew file: kNN-TL/examples/criss/save_encoder.py\n\tnew file: kNN-TL/examples/criss/sentence_retrieval/encoder_analysis.py\n\tnew file: kNN-TL/examples/criss/sentence_retrieval/sentence_retrieval_tatoeba.sh\n\tnew file: kNN-TL/examples/criss/unsupervised_mt/eval.sh\n\tnew file: kNN-TL/examples/cross_lingual_language_model/README.md\n\tnew file: kNN-TL/examples/data2vec/README.md\n\tnew file: kNN-TL/examples/data2vec/config/audio/pretraining/base_librispeech.yaml\n\tnew file: kNN-TL/examples/data2vec/config/text/pretraining/base.yaml\n\tnew file: kNN-TL/examples/data2vec/models/data2vec_audio.py\n\tnew file: kNN-TL/examples/data2vec/models/data2vec_text.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/README.md\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/__init__.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/config/deen.yaml\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/criterions/__init__.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/criterions/discriminative_reranking_criterion.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/drnmt_rerank.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/models/__init__.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/models/discriminative_reranking_model.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/scripts/prep_data.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/tasks/__init__.py\n\tnew file: kNN-TL/examples/discriminative_reranking_nmt/tasks/discriminative_reranking_task.py\n\tnew file: kNN-TL/examples/fast_noisy_channel/README.md\n\tnew file: kNN-TL/examples/fast_noisy_channel/__init__.py\n\tnew file: kNN-TL/examples/fast_noisy_channel/noisy_channel_beam_search.py\n\tnew file: kNN-TL/examples/fast_noisy_channel/noisy_channel_sequence_generator.py\n\tnew file: kNN-TL/examples/fast_noisy_channel/noisy_channel_translation.py\n\tnew file: kNN-TL/examples/flores101/README.md\n\tnew file: kNN-TL/examples/flores101/flores_logo.png\n\tnew file: kNN-TL/examples/fully_sharded_data_parallel/README.md\n\tnew file: kNN-TL/examples/gottbert/README.md\n\tnew file: kNN-TL/examples/hubert/README.md\n\tnew file: kNN-TL/examples/hubert/config/decode/ax_sweep/ngram.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/ax_sweep/transformer.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/infer_fsqlm.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/infer_kenlm.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/infer_viterbi.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/run/submitit_slurm.yaml\n\tnew file: kNN-TL/examples/hubert/config/decode/run/submitit_slurm_8gpu.yaml\n\tnew file: kNN-TL/examples/hubert/config/finetune/base_10h.yaml\n\tnew file: kNN-TL/examples/hubert/config/finetune/ckpt/it1.yaml\n\tnew file: kNN-TL/examples/hubert/config/finetune/lm/ls_4gram.yaml\n\tnew file: kNN-TL/examples/hubert/config/finetune/run/submitit_reg.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/data/iter1.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/data/iter2.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/hubert_base_librispeech.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/hubert_large_librivox.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/hubert_xlarge_librivox.yaml\n\tnew file: kNN-TL/examples/hubert/config/pretrain/run/submitit_reg.yaml\n\tnew file: kNN-TL/examples/hubert/measure_teacher_quality.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/README.md\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/dump_hubert_feature.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/dump_hubert_feature_s2t.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/dump_km_label.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/dump_mfcc_feature.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/dump_w2v2_feature.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/feature_utils.py\n\tnew file: kNN-TL/examples/hubert/simple_kmeans/learn_kmeans.py\n\tnew file: kNN-TL/examples/hubert/tests/6313-76958-0021.flac\n\tnew file: kNN-TL/examples/hubert/tests/sample.base.L9.km500.km\n\tnew file: kNN-TL/examples/hubert/tests/sample.base.L9.len\n\tnew file: kNN-TL/examples/hubert/tests/sample.base.L9.npy\n\tnew file: kNN-TL/examples/hubert/tests/sample.large.L20.len\n\tnew file: kNN-TL/examples/hubert/tests/sample.large.L20.npy\n\tnew file: kNN-TL/examples/hubert/tests/sample.large.hypo.word\n\tnew file: kNN-TL/examples/hubert/tests/sample.xlarge.L30.len\n\tnew file: kNN-TL/examples/hubert/tests/sample.xlarge.L30.npy\n\tnew file: kNN-TL/examples/hubert/tests/sample.xlarge.hypo.word\n\tnew file: kNN-TL/examples/hubert/tests/test_feature_and_unit.sh\n\tnew file: kNN-TL/examples/hubert/tests/test_finetuned_asr.sh\n\tnew file: kNN-TL/examples/hubert/update_ckpt.py\n\tnew file: kNN-TL/examples/joint_alignment_translation/README.md\n\tnew file: kNN-TL/examples/joint_alignment_translation/prepare-wmt18en2de_no_norm_no_escape_no_agressive.sh\n\tnew file: kNN-TL/examples/language_model/README.adaptive_inputs.md\n\tnew file: kNN-TL/examples/language_model/README.conv.md\n\tnew file: kNN-TL/examples/language_model/README.md\n\tnew file: kNN-TL/examples/language_model/prepare-wikitext-103.sh\n\tnew file: kNN-TL/examples/laser/README.md\n\tnew file: kNN-TL/examples/laser/laser_src/__init__.py\n\tnew file: kNN-TL/examples/laser/laser_src/laser_lstm.py\n\tnew file: kNN-TL/examples/laser/laser_src/laser_task.py\n\tnew file: kNN-TL/examples/laser/laser_src/laser_transformer.py\n\tnew file: kNN-TL/examples/laser/laser_src/multitask_data_utils.py\n\tnew file: kNN-TL/examples/latent_depth/README.md\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/__init__.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/loss/__init__.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/loss/latent_depth.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/models/__init__.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/models/latent_multilingual_transformer.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/models/latent_transformer.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/modules/__init__.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/modules/latent_layers.py\n\tnew file: kNN-TL/examples/latent_depth/latent_depth_src/multilingual_translation_latent_depth.py\n\tnew file: kNN-TL/examples/layerdrop/README.md\n\tnew file: kNN-TL/examples/linformer/README.md\n\tnew file: kNN-TL/examples/linformer/linformer_src/__init__.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/models/__init__.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/models/linformer_roberta.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/modules/__init__.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/modules/linformer_sentence_encoder.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/modules/linformer_sentence_encoder_layer.py\n\tnew file: kNN-TL/examples/linformer/linformer_src/modules/multihead_linear_attention.py\n\tnew file: kNN-TL/examples/m2m_100/README.md\n\tnew file: kNN-TL/examples/m2m_100/install_dependecies.sh\n\tnew file: kNN-TL/examples/m2m_100/process_data/clean_histogram.py\n\tnew file: kNN-TL/examples/m2m_100/process_data/dedup_data.py\n\tnew file: kNN-TL/examples/m2m_100/process_data/remove_too_much_punc.py\n\tnew file: kNN-TL/examples/m2m_100/tok.sh\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/README.md\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/seg_ja.sh\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/seg_ko.sh\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/thirdparty/.gitignore\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/tokenize_indic.py\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/tokenize_thai.py\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/tokenize_zh.py\n\tnew file: kNN-TL/examples/m2m_100/tokenizers/tokenizer_ar.sh\n\tnew file: kNN-TL/examples/mbart/README.md\n\tnew file: kNN-TL/examples/megatron_11b/README.md\n\tnew file: kNN-TL/examples/megatron_11b/detok.py\n\tnew file: kNN-TL/examples/moe_lm/README.md\n\tnew file: kNN-TL/examples/moe_lm/data_card.md\n\tnew file: kNN-TL/examples/moe_lm/model_card.md\n\tnew file: kNN-TL/examples/multilingual/ML50_langs.txt\n\tnew file: kNN-TL/examples/multilingual/README.md\n\tnew file: kNN-TL/examples/multilingual/data_scripts/README.md\n\tnew file: kNN-TL/examples/multilingual/data_scripts/binarize.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/check_iswlt_test_data.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/check_self_overlaps.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/check_valid_test_overlaps.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/dedup_all.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_ML50_v1.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_af_xh.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_flores_data.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_iitb.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_iwslt_and_extract.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_lotus.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_ted_and_extract.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_wat19_my.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_wmt19_and_before.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/download_wmt20.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/preprocess_ML50_v1.sh\n\tnew file: kNN-TL/examples/multilingual/data_scripts/remove_valid_test_in_train.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/requirement.txt\n\tnew file: kNN-TL/examples/multilingual/data_scripts/utils/dedup.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/utils/fasttext_multi_filter.py\n\tnew file: kNN-TL/examples/multilingual/data_scripts/utils/strip_sgm.sh\n\tnew file: kNN-TL/examples/multilingual/finetune_multilingual_model.sh\n\tnew file: kNN-TL/examples/multilingual/multilingual_fairseq_gen.sh\n\tnew file: kNN-TL/examples/multilingual/train_multilingual_model.sh\n\tnew file: kNN-TL/examples/noisychannel/README.md\n\tnew file: kNN-TL/examples/noisychannel/__init__.py\n\tnew file: kNN-TL/examples/noisychannel/rerank.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_generate.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_options.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_score_bw.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_score_lm.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_tune.py\n\tnew file: kNN-TL/examples/noisychannel/rerank_utils.py\n\tnew file: kNN-TL/examples/nonautoregressive_translation/README.md\n\tnew file: kNN-TL/examples/nonautoregressive_translation/scripts.md\n\tnew file: kNN-TL/examples/normformer/README.md\n\tnew file: kNN-TL/examples/normformer/train_lm.sh\n\tnew file: kNN-TL/examples/operators/alignment_train_cpu.cpp\n\tnew file: kNN-TL/examples/operators/alignment_train_cuda.cpp\n\tnew file: kNN-TL/examples/operators/alignment_train_cuda.h\n\tnew file: kNN-TL/examples/operators/alignment_train_kernel.cu\n\tnew file: kNN-TL/examples/operators/utils.h\n\tnew file: kNN-TL/examples/paraphraser/README.md\n\tnew file: kNN-TL/examples/paraphraser/paraphrase.py\n\tnew file: kNN-TL/examples/pay_less_attention_paper/README.md\n\tnew file: kNN-TL/examples/pointer_generator/README.md\n\tnew file: kNN-TL/examples/pointer_generator/README.xsum.md\n\tnew file: kNN-TL/examples/pointer_generator/pointer_generator_src/__init__.py\n\tnew file: kNN-TL/examples/pointer_generator/pointer_generator_src/transformer_pg.py\n\tnew file: kNN-TL/examples/pointer_generator/postprocess.py\n\tnew file: kNN-TL/examples/pointer_generator/preprocess.py\n\tnew file: kNN-TL/examples/quant_noise/README.md\n\tnew file: kNN-TL/examples/quant_noise/transformer_quantization_config.yaml\n\tnew file: kNN-TL/examples/roberta/README.custom_classification.md\n\tnew file: kNN-TL/examples/roberta/README.glue.md\n\tnew file: kNN-TL/examples/roberta/README.md\n\tnew file: kNN-TL/examples/roberta/README.pretraining.md\n\tnew file: kNN-TL/examples/roberta/README.race.md\n\tnew file: kNN-TL/examples/roberta/commonsense_qa/README.md\n\tnew file: kNN-TL/examples/roberta/commonsense_qa/__init__.py\n\tnew file: kNN-TL/examples/roberta/commonsense_qa/commonsense_qa_task.py\n\tnew file: kNN-TL/examples/roberta/commonsense_qa/download_cqa_data.sh\n\tnew file: kNN-TL/examples/roberta/config/finetuning/cola.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/mnli.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/mrpc.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/qnli.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/qqp.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/rte.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/sst_2.yaml\n\tnew file: kNN-TL/examples/roberta/config/finetuning/sts_b.yaml\n\tnew file: kNN-TL/examples/roberta/config/pretraining/base.yaml\n\tnew file: kNN-TL/examples/roberta/multiprocessing_bpe_encoder.py\n\tnew file: kNN-TL/examples/roberta/preprocess_GLUE_tasks.sh\n\tnew file: kNN-TL/examples/roberta/preprocess_RACE.py\n\tnew file: kNN-TL/examples/roberta/preprocess_RACE.sh\n\tnew file: kNN-TL/examples/roberta/wsc/README.md\n\tnew file: kNN-TL/examples/roberta/wsc/__init__.py\n\tnew file: kNN-TL/examples/roberta/wsc/wsc_criterion.py\n\tnew file: kNN-TL/examples/roberta/wsc/wsc_task.py\n\tnew file: kNN-TL/examples/roberta/wsc/wsc_utils.py\n\tnew file: kNN-TL/examples/rxf/README.md\n\tnew file: kNN-TL/examples/rxf/__init__.py\n\tnew file: kNN-TL/examples/rxf/rxf_src/__init__.py\n\tnew file: kNN-TL/examples/rxf/rxf_src/label_smoothed_cross_entropy_r3f.py\n\tnew file: kNN-TL/examples/rxf/rxf_src/sentence_prediction_r3f.py\n\tnew file: kNN-TL/examples/scaling_nmt/README.md\n\tnew file: kNN-TL/examples/shuffled_word_order/README.finetuning.md\n\tnew file: kNN-TL/examples/shuffled_word_order/README.md\n\tnew file: kNN-TL/examples/simultaneous_translation/README.md\n\tnew file: kNN-TL/examples/simultaneous_translation/__init__.py\n\tnew file: kNN-TL/examples/simultaneous_translation/docs/ende-mma.md\n\tnew file: kNN-TL/examples/simultaneous_translation/docs/enja-waitk.md\n\tnew file: kNN-TL/examples/simultaneous_translation/eval/agents/simul_t2t_enja.py\n\tnew file: kNN-TL/examples/simultaneous_translation/models/__init__.py\n\tnew file: kNN-TL/examples/simultaneous_translation/models/convtransformer_simul_trans.py\n\tnew file: kNN-TL/examples/simultaneous_translation/models/transformer_monotonic_attention.py\n\tnew file: kNN-TL/examples/simultaneous_translation/modules/__init__.py\n\tnew file: kNN-TL/examples/simultaneous_translation/modules/fixed_pre_decision.py\n\tnew file: kNN-TL/examples/simultaneous_translation/modules/monotonic_multihead_attention.py\n\tnew file: kNN-TL/examples/simultaneous_translation/modules/monotonic_transformer_layer.py\n\tnew file: kNN-TL/examples/simultaneous_translation/tests/test_alignment_train.py\n\tnew file: kNN-TL/examples/simultaneous_translation/tests/test_text_models.py\n\tnew file: kNN-TL/examples/simultaneous_translation/utils/__init__.py\n\tnew file: kNN-TL/examples/simultaneous_translation/utils/functions.py\n\tnew file: kNN-TL/examples/simultaneous_translation/utils/monotonic_attention.py\n\tnew file: kNN-TL/examples/simultaneous_translation/utils/p_choose_strategy.py\n\tnew file: kNN-TL/examples/speech_recognition/README.md\n\tnew file: kNN-TL/examples/speech_recognition/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/criterions/ASG_loss.py\n\tnew file: kNN-TL/examples/speech_recognition/criterions/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/criterions/cross_entropy_acc.py\n\tnew file: kNN-TL/examples/speech_recognition/data/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/data/asr_dataset.py\n\tnew file: kNN-TL/examples/speech_recognition/data/collaters.py\n\tnew file: kNN-TL/examples/speech_recognition/data/data_utils.py\n\tnew file: kNN-TL/examples/speech_recognition/data/replabels.py\n\tnew file: kNN-TL/examples/speech_recognition/datasets/asr_prep_json.py\n\tnew file: kNN-TL/examples/speech_recognition/datasets/prepare-librispeech.sh\n\tnew file: kNN-TL/examples/speech_recognition/infer.py\n\tnew file: kNN-TL/examples/speech_recognition/kaldi/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/kaldi/add-self-loop-simple.cc\n\tnew file: kNN-TL/examples/speech_recognition/kaldi/config/kaldi_initializer.yaml\n\tnew file: kNN-TL/examples/speech_recognition/kaldi/kaldi_decoder.py\n\tnew file: kNN-TL/examples/speech_recognition/kaldi/kaldi_initializer.py\n\tnew file: kNN-TL/examples/speech_recognition/models/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/models/vggtransformer.py\n\tnew file: kNN-TL/examples/speech_recognition/models/w2l_conv_glu_enc.py\n\tnew file: kNN-TL/examples/speech_recognition/new/README.md\n\tnew file: kNN-TL/examples/speech_recognition/new/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/new/conf/hydra/sweeper/ax.yaml\n\tnew file: kNN-TL/examples/speech_recognition/new/conf/infer.yaml\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/base_decoder.py\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/decoder.py\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/decoder_config.py\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/flashlight_decoder.py\n\tnew file: kNN-TL/examples/speech_recognition/new/decoders/viterbi_decoder.py\n\tnew file: kNN-TL/examples/speech_recognition/new/infer.py\n\tnew file: kNN-TL/examples/speech_recognition/tasks/__init__.py\n\tnew file: kNN-TL/examples/speech_recognition/tasks/speech_recognition.py\n\tnew file: kNN-TL/examples/speech_recognition/utils/wer_utils.py\n\tnew file: kNN-TL/examples/speech_recognition/w2l_decoder.py\n\tnew file: kNN-TL/examples/speech_synthesis/README.md\n\tnew file: kNN-TL/examples/speech_synthesis/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/data_utils.py\n\tnew file: kNN-TL/examples/speech_synthesis/docs/common_voice_example.md\n\tnew file: kNN-TL/examples/speech_synthesis/docs/ljspeech_example.md\n\tnew file: kNN-TL/examples/speech_synthesis/docs/vctk_example.md\n\tnew file: kNN-TL/examples/speech_synthesis/evaluation/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/evaluation/eval_asr.py\n\tnew file: kNN-TL/examples/speech_synthesis/evaluation/eval_f0.py\n\tnew file: kNN-TL/examples/speech_synthesis/evaluation/eval_sp.py\n\tnew file: kNN-TL/examples/speech_synthesis/evaluation/get_eval_manifest.py\n\tnew file: kNN-TL/examples/speech_synthesis/generate_waveform.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoise_and_vad_audio.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoiser/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoiser/demucs.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoiser/pretrained.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoiser/resample.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/denoiser/utils.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/get_common_voice_audio_manifest.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/get_feature_manifest.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/get_ljspeech_audio_manifest.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/get_speaker_embedding.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/get_vctk_audio_manifest.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/speaker_embedder/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/preprocessing/vad/__init__.py\n\tnew file: kNN-TL/examples/speech_synthesis/utils.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/README.md\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/__init__.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/configs/mustc_noise.list\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/criterions/__init__.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/criterions/multi_modality_compound.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/criterions/multi_modality_cross_entropy.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/criterions/text_guide_cross_entropy_acc.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/data/pair_denoising_dataset.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/docs/ende-mustc.md\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/docs/iwslt2021.md\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/docs/pre-training.md\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/models/__init__.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/models/joint_speech_text_pretrain_transformer.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/models/s2t_dualinputtransformer.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/models/s2t_dualinputwavtransformer.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/models/s2t_dualinputxmtransformer.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/scripts/convert_model.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/scripts/g2p_encode.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/tasks/__init__.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/tasks/pair_denoising.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/tasks/speech_text_denoise_pretrain.py\n\tnew file: kNN-TL/examples/speech_text_joint_to_text/tasks/speech_text_joint.py\n\tnew file: kNN-TL/examples/speech_to_speech/README.md\n\tnew file: kNN-TL/examples/speech_to_speech/__init__.py\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/README.md\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/configs/2StageS2ST.yaml\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/configs/3StageS2ST.yaml\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/configs/DirectS2U.yaml\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/configs/S2T.yaml\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/core.py\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/data_utils.py\n\tnew file: kNN-TL/examples/speech_to_speech/benchmarking/get_metrics.py\n\tnew file: kNN-TL/examples/speech_to_speech/docs/direct_s2st_discrete_units.md\n\tnew file: kNN-TL/examples/speech_to_speech/docs/enhanced_direct_s2st_discrete_units.md\n\tnew file: kNN-TL/examples/speech_to_speech/docs/textless_s2st_real_data.md\n\tnew file: kNN-TL/examples/speech_to_speech/generate_waveform_from_code.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/__init__.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/data_utils.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/prep_s2spect_data.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/prep_s2ut_data.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/prep_sn_data.py\n\tnew file: kNN-TL/examples/speech_to_speech/preprocessing/prep_sn_output_data.py\n\tnew file: kNN-TL/examples/speech_to_text/README.md\n\tnew file: kNN-TL/examples/speech_to_text/data_utils.py\n\tnew file: kNN-TL/examples/speech_to_text/docs/covost_example.md\n\tnew file: kNN-TL/examples/speech_to_text/docs/librispeech_example.md\n\tnew file: kNN-TL/examples/speech_to_text/docs/mtedx_example.md\n\tnew file: kNN-TL/examples/speech_to_text/docs/mustc_example.md\n\tnew file: kNN-TL/examples/speech_to_text/docs/simulst_mustc_example.md\n\tnew file: kNN-TL/examples/speech_to_text/prep_covost_data.py\n\tnew file: kNN-TL/examples/speech_to_text/prep_librispeech_data.py\n\tnew file: kNN-TL/examples/speech_to_text/prep_mtedx_data.py\n\tnew file: kNN-TL/examples/speech_to_text/prep_mustc_data.py\n\tnew file: kNN-TL/examples/speech_to_text/seg_mustc_data.py\n\tnew file: kNN-TL/examples/speech_to_text/simultaneous_translation/agents/fairseq_simul_st_agent.py\n\tnew file: kNN-TL/examples/stories/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/abx_metrics/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/abx_metrics/dump_abx_feats.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/continuation_eval.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/misc/bleu_utils.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/misc/cut_as.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/misc/dict.ltr.txt\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/ppx.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/metrics/asr_metrics/self_auto_bleu.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/__init__.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/clustering/__init__.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/clustering/cluster_kmeans.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/clustering/dump_feats.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/clustering/quantize_with_kmeans.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/clustering/utils.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/pretrained/cpc_feature_reader.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/pretrained/hubert_feature_reader.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/pretrained/logmel_feature_reader.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/pretrained/utils.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/speech2unit/pretrained/w2v2_feature_reader.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/tools/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/tools/resynthesize_speech.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/ulm/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/ulm/sample.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/README.md\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/convert_to_16k.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/glow.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/multiproc.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/synthesize_audio_from_units.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/__init__.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/audio_processing.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/cleaners.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/cmudict.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/layers.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/model.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/numbers.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/stft.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/symbols.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/text.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/utils.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tacotron2/waveglow_denoiser.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/tts_data.py\n\tnew file: kNN-TL/examples/textless_nlp/gslm/unit2speech/utils.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/README.md\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/data_utils.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/eval/__init__.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/eval/cont_metrics.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/generate_waveform.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/inference_dataset.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/naive_decoder.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/prepare_dataset.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/preprocess_f0.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/quantize_f0.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/sample/__init__.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/sample/sample.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/scripts/join_units_manifest.py\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/scripts/prepare_data.sh\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/scripts/prepare_f0_quantization.sh\n\tnew file: kNN-TL/examples/textless_nlp/pgslm/truncated_laplace.py\n\tnew file: kNN-TL/examples/textless_nlp/speech-resynth/README.md\n\tnew file: kNN-TL/examples/textless_nlp/speech-resynth/img/fig.png\n\tnew file: kNN-TL/examples/translation/README.md\n\tnew file: kNN-TL/examples/translation/prepare-iwslt14.sh\n\tnew file: kNN-TL/examples/translation/prepare-iwslt17-multilingual.sh\n\tnew file: kNN-TL/examples/translation/prepare-wmt14en2de.sh\n\tnew file: kNN-TL/examples/translation/prepare-wmt14en2fr.sh\n\tnew file: kNN-TL/examples/translation_moe/README.md\n\tnew file: kNN-TL/examples/translation_moe/score.py\n\tnew file: kNN-TL/examples/translation_moe/translation_moe_src/__init__.py\n\tnew file: kNN-TL/examples/translation_moe/translation_moe_src/logsumexp_moe.py\n\tnew file: kNN-TL/examples/translation_moe/translation_moe_src/mean_pool_gating_network.py\n\tnew file: kNN-TL/examples/translation_moe/translation_moe_src/translation_moe.py\n\tnew file: kNN-TL/examples/truncated_bptt/README.md\n\tnew file: kNN-TL/examples/truncated_bptt/__init__.py\n\tnew file: kNN-TL/examples/truncated_bptt/transformer_xl_model.py\n\tnew file: kNN-TL/examples/truncated_bptt/truncated_bptt_lm_task.py\n\tnew file: kNN-TL/examples/unsupervised_quality_estimation/README.md\n\tnew file: kNN-TL/examples/unsupervised_quality_estimation/aggregate_scores.py\n\tnew file: kNN-TL/examples/unsupervised_quality_estimation/meteor.py\n\tnew file: kNN-TL/examples/unsupervised_quality_estimation/repeat_lines.py\n\tnew file: kNN-TL/examples/wav2vec/README.md\n\tnew file: kNN-TL/examples/wav2vec/__init__.py\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/base_100h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/base_10h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/base_10m.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/base_1h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/base_960h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/vox_100h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/vox_10h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/vox_10m.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/vox_1h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/finetuning/vox_960h.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_base_librispeech.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_conformer_base_librispeech.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_conformer_large_librivox.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_large_librivox.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu-pod.yaml\n\tnew file: kNN-TL/examples/wav2vec/config/pretraining/wav2vec2_large_librivox_tpu.yaml\n\tnew file: kNN-TL/examples/wav2vec/libri_labels.py\n\tnew file: kNN-TL/examples/wav2vec/scripts/binarize_manifest.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/README.md\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/__init__.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/finetuning/w2v_finetune.yaml\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/gan/w2vu.yaml\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/generate/viterbi.yaml\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_matched/test.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_matched/train.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_matched/train_text.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_matched/valid.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_unmatched/test.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_unmatched/train.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_unmatched/train_text.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/config/timit_unmatched/valid.uid\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/data/__init__.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/data/extracted_features_dataset.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/data/random_input_dataset.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/README.md\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/cmd.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_phone.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step1.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/decode_word_step2.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/copy_aligned_text.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/decode.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_data_from_w2v.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lang_word.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/prepare_lm.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/score.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/show_wer.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/train_subset_lgbeam.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/local/unsup_select_decode_word.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/path.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_deltas.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_lda_mllt.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/steps_gan/train_sat.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/kaldi_self_train/st/train.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/models/__init__.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/models/wav2vec_u.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/apply_pca.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/copy_labels.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/filter_lexicon.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/filter_tsv.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/g2p_wrd_to_phn.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/ltr_to_wrd.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/mean_pool.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/merge_clusters.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/normalize_and_filter_text.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/normalize_text.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/pca.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/phonemize_with_sil.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/prepare_audio.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/prepare_text.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/prepare_timit.sh\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/remove_silence.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/vads.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/wav2vec_apply_cluster_faiss.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/wav2vec_cluster_faiss.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/wav2vec_extract_features.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/wer.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/scripts/wrd_to_ltr.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/tasks/__init__.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/tasks/unpaired_audio_text.py\n\tnew file: kNN-TL/examples/wav2vec/unsupervised/w2vu_generate.py\n\tnew file: kNN-TL/examples/wav2vec/vq-wav2vec_featurize.py\n\tnew file: kNN-TL/examples/wav2vec/wav2vec_featurize.py\n\tnew file: kNN-TL/examples/wav2vec/wav2vec_manifest.py\n\tnew file: kNN-TL/examples/wav2vec/xlsr/README.md\n\tnew file: kNN-TL/examples/wav2vec/xlsr/config/finetune.yaml\n\tnew file: kNN-TL/examples/wmt19/README.md\n\tnew file: kNN-TL/examples/wmt20/README.md\n\tnew file: kNN-TL/examples/wmt21/README.md\n\tnew file: kNN-TL/examples/wmt21/eval.sh\n\tnew file: kNN-TL/examples/wmt21/scripts/normalize-punctuation.perl\n\tnew file: kNN-TL/examples/wmt21/scripts/replace-unicode-punctuation.perl\n\tnew file: kNN-TL/examples/womens_bios/README.md\n\tnew file: kNN-TL/examples/womens_bios/query_occupations_from_wikidata.py\n\tnew file: kNN-TL/examples/xformers/README.md\n\tnew file: kNN-TL/examples/xglm/README.md\n\tnew file: kNN-TL/examples/xglm/model_card.md\n\tnew file: kNN-TL/examples/xlmr/README.md\n\tnew file: kNN-TL/examples/xmod/README.md\n\tnew file: kNN-TL/examples/xmod/preprocess_nli.py\n\tnew file: kNN-TL/fairseq/__init__.py\n\tnew file: kNN-TL/fairseq/benchmark/__init__.py\n\tnew file: kNN-TL/fairseq/benchmark/benchmark_multihead_attention.py\n\tnew file: kNN-TL/fairseq/benchmark/dummy_dataset.py\n\tnew file: kNN-TL/fairseq/benchmark/dummy_lm.py\n\tnew file: kNN-TL/fairseq/benchmark/dummy_masked_lm.py\n\tnew file: kNN-TL/fairseq/benchmark/dummy_model.py\n\tnew file: kNN-TL/fairseq/benchmark/dummy_mt.py\n\tnew file: kNN-TL/fairseq/binarizer.py\n\tnew file: kNN-TL/fairseq/checkpoint_utils.py\n\tnew file: kNN-TL/fairseq/clib/cuda/ngram_repeat_block_cuda.cpp\n\tnew file: kNN-TL/fairseq/clib/cuda/ngram_repeat_block_cuda_kernel.cu\n\tnew file: kNN-TL/fairseq/clib/libbase/balanced_assignment.cpp\n\tnew file: kNN-TL/fairseq/clib/libbleu/libbleu.cpp\n\tnew file: kNN-TL/fairseq/clib/libbleu/module.cpp\n\tnew file: kNN-TL/fairseq/clib/libnat/edit_dist.cpp\n\tnew file: kNN-TL/fairseq/clib/libnat_cuda/binding.cpp\n\tnew file: kNN-TL/fairseq/clib/libnat_cuda/edit_dist.cu\n\tnew file: kNN-TL/fairseq/clib/libnat_cuda/edit_dist.h\n\tnew file: kNN-TL/fairseq/config/__init__.py\n\tnew file: kNN-TL/fairseq/config/config.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_baevski_gbw.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_baevski_wiki103.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_big.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_gbw.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_gpt.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_gpt2_big.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_gpt2_medium.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_gpt2_small.yaml\n\tnew file: kNN-TL/fairseq/config/model/transformer_lm/transformer_lm_wiki103.yaml\n\tnew file: kNN-TL/fairseq/config/model/wav2vec/vq_wav2vec_gumbel.yaml\n\tnew file: kNN-TL/fairseq/config/model/wav2vec2/wav2vec2_base.yaml\n\tnew file: kNN-TL/fairseq/config/model/wav2vec2/wav2vec2_large.yaml\n\tnew file: kNN-TL/fairseq/criterions/__init__.py\n\tnew file: kNN-TL/fairseq/criterions/adaptive_loss.py\n\tnew file: kNN-TL/fairseq/criterions/composite_loss.py\n\tnew file: kNN-TL/fairseq/criterions/cross_entropy.py\n\tnew file: kNN-TL/fairseq/criterions/ctc.py\n\tnew file: kNN-TL/fairseq/criterions/fairseq_criterion.py\n\tnew file: kNN-TL/fairseq/criterions/fastspeech2_loss.py\n\tnew file: kNN-TL/fairseq/criterions/hubert_criterion.py\n\tnew file: kNN-TL/fairseq/criterions/label_smoothed_cross_entropy.py\n\tnew file: kNN-TL/fairseq/criterions/label_smoothed_cross_entropy_latency_augmented.py\n\tnew file: kNN-TL/fairseq/criterions/label_smoothed_cross_entropy_with_alignment.py\n\tnew file: kNN-TL/fairseq/criterions/label_smoothed_cross_entropy_with_ctc.py\n\tnew file: kNN-TL/fairseq/criterions/label_smoothed_cross_entropy_xkd.py\n\tnew file: kNN-TL/fairseq/criterions/legacy_masked_lm.py\n\tnew file: kNN-TL/fairseq/criterions/masked_lm.py\n\tnew file: kNN-TL/fairseq/criterions/model_criterion.py\n\tnew file: kNN-TL/fairseq/criterions/nat_loss.py\n\tnew file: kNN-TL/fairseq/criterions/sentence_prediction.py\n\tnew file: kNN-TL/fairseq/criterions/sentence_prediction_adapters.py\n\tnew file: kNN-TL/fairseq/criterions/sentence_ranking.py\n\tnew file: kNN-TL/fairseq/criterions/speech_to_speech_criterion.py\n\tnew file: kNN-TL/fairseq/criterions/speech_ulm_criterion.py\n\tnew file: kNN-TL/fairseq/criterions/tacotron2_loss.py\n\tnew file: kNN-TL/fairseq/criterions/wav2vec_criterion.py\n\tnew file: kNN-TL/fairseq/data/__init__.py\n\tnew file: kNN-TL/fairseq/data/add_target_dataset.py\n\tnew file: kNN-TL/fairseq/data/append_token_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/__init__.py\n\tnew file: kNN-TL/fairseq/data/audio/audio_utils.py\n\tnew file: kNN-TL/fairseq/data/audio/data_cfg.py\n\tnew file: kNN-TL/fairseq/data/audio/feature_transforms/__init__.py\n\tnew file: kNN-TL/fairseq/data/audio/feature_transforms/delta_deltas.py\n\tnew file: kNN-TL/fairseq/data/audio/feature_transforms/global_cmvn.py\n\tnew file: kNN-TL/fairseq/data/audio/feature_transforms/specaugment.py\n\tnew file: kNN-TL/fairseq/data/audio/feature_transforms/utterance_cmvn.py\n\tnew file: kNN-TL/fairseq/data/audio/frm_text_to_speech_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/hubert_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/multi_modality_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/raw_audio_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/speech_to_speech_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/speech_to_text_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/speech_to_text_joint_dataset.py\n\tnew file: kNN-TL/fairseq/data/audio/text_to_speech_dataset.py\n\tnew file: kNN-TL/fairseq/data/backtranslation_dataset.py\n\tnew file: kNN-TL/fairseq/data/base_wrapper_dataset.py\n\tnew file: kNN-TL/fairseq/data/bucket_pad_length_dataset.py\n\tnew file: kNN-TL/fairseq/data/codedataset.py\n\tnew file: kNN-TL/fairseq/data/colorize_dataset.py\n\tnew file: kNN-TL/fairseq/data/concat_dataset.py\n\tnew file: kNN-TL/fairseq/data/concat_sentences_dataset.py\n\tnew file: kNN-TL/fairseq/data/data_utils.py\n\tnew file: kNN-TL/fairseq/data/data_utils_fast.pyx\n\tnew file: kNN-TL/fairseq/data/denoising_dataset.py\n\tnew file: kNN-TL/fairseq/data/dictionary.py\n\tnew file: kNN-TL/fairseq/data/encoders/__init__.py\n\tnew file: kNN-TL/fairseq/data/encoders/byte_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/byte_utils.py\n\tnew file: kNN-TL/fairseq/data/encoders/bytes.py\n\tnew file: kNN-TL/fairseq/data/encoders/characters.py\n\tnew file: kNN-TL/fairseq/data/encoders/fastbpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/gpt2_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/gpt2_bpe_utils.py\n\tnew file: kNN-TL/fairseq/data/encoders/hf_bert_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/hf_byte_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/moses_tokenizer.py\n\tnew file: kNN-TL/fairseq/data/encoders/nltk_tokenizer.py\n\tnew file: kNN-TL/fairseq/data/encoders/sentencepiece_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/space_tokenizer.py\n\tnew file: kNN-TL/fairseq/data/encoders/subword_nmt_bpe.py\n\tnew file: kNN-TL/fairseq/data/encoders/utils.py\n\tnew file: kNN-TL/fairseq/data/fairseq_dataset.py\n\tnew file: kNN-TL/fairseq/data/fasta_dataset.py\n\tnew file: kNN-TL/fairseq/data/huffman/__init__.py\n\tnew file: kNN-TL/fairseq/data/huffman/huffman_coder.py\n\tnew file: kNN-TL/fairseq/data/huffman/huffman_mmap_indexed_dataset.py\n\tnew file: kNN-TL/fairseq/data/id_dataset.py\n\tnew file: kNN-TL/fairseq/data/indexed_dataset.py\n\tnew file: kNN-TL/fairseq/data/iterators.py\n\tnew file: kNN-TL/fairseq/data/language_pair_dataset.py\n\tnew file: kNN-TL/fairseq/data/legacy/__init__.py\n\tnew file: kNN-TL/fairseq/data/legacy/block_pair_dataset.py\n\tnew file: kNN-TL/fairseq/data/legacy/masked_lm_dataset.py\n\tnew file: kNN-TL/fairseq/data/legacy/masked_lm_dictionary.py\n\tnew file: kNN-TL/fairseq/data/list_dataset.py\n\tnew file: kNN-TL/fairseq/data/lm_context_window_dataset.py\n\tnew file: kNN-TL/fairseq/data/lru_cache_dataset.py\n\tnew file: kNN-TL/fairseq/data/mask_tokens_dataset.py\n\tnew file: kNN-TL/fairseq/data/monolingual_dataset.py\n\tnew file: kNN-TL/fairseq/data/multi_corpus_dataset.py\n\tnew file: kNN-TL/fairseq/data/multi_corpus_sampled_dataset.py\n\tnew file: kNN-TL/fairseq/data/multilingual/__init__.py\n\tnew file: kNN-TL/fairseq/data/multilingual/multilingual_data_manager.py\n\tnew file: kNN-TL/fairseq/data/multilingual/multilingual_utils.py\n\tnew file: kNN-TL/fairseq/data/multilingual/sampled_multi_dataset.py\n\tnew file: kNN-TL/fairseq/data/multilingual/sampled_multi_epoch_dataset.py\n\tnew file: kNN-TL/fairseq/data/multilingual/sampling_method.py\n\tnew file: kNN-TL/fairseq/data/nested_dictionary_dataset.py\n\tnew file: kNN-TL/fairseq/data/noising.py\n\tnew file: kNN-TL/fairseq/data/num_samples_dataset.py\n\tnew file: kNN-TL/fairseq/data/numel_dataset.py\n\tnew file: kNN-TL/fairseq/data/offset_tokens_dataset.py\n\tnew file: kNN-TL/fairseq/data/pad_dataset.py\n\tnew file: kNN-TL/fairseq/data/plasma_utils.py\n\tnew file: kNN-TL/fairseq/data/prepend_dataset.py\n\tnew file: kNN-TL/fairseq/data/prepend_token_dataset.py\n\tnew file: kNN-TL/fairseq/data/raw_label_dataset.py\n\tnew file: kNN-TL/fairseq/data/replace_dataset.py\n\tnew file: kNN-TL/fairseq/data/resampling_dataset.py\n\tnew file: kNN-TL/fairseq/data/roll_dataset.py\n\tnew file: kNN-TL/fairseq/data/round_robin_zip_datasets.py\n\tnew file: kNN-TL/fairseq/data/shorten_dataset.py\n\tnew file: kNN-TL/fairseq/data/sort_dataset.py\n\tnew file: kNN-TL/fairseq/data/strip_token_dataset.py\n\tnew file: kNN-TL/fairseq/data/subsample_dataset.py\n\tnew file: kNN-TL/fairseq/data/text_compressor.py\n\tnew file: kNN-TL/fairseq/data/token_block_dataset.py\n\tnew file: kNN-TL/fairseq/data/token_block_utils_fast.pyx\n\tnew file: kNN-TL/fairseq/data/transform_eos_concat_langpair_dataset.py\n\tnew file: kNN-TL/fairseq/data/transform_eos_dataset.py\n\tnew file: kNN-TL/fairseq/data/transform_eos_lang_pair_dataset.py\n\tnew file: kNN-TL/fairseq/dataclass/__init__.py\n\tnew file: kNN-TL/fairseq/dataclass/configs.py\n\tnew file: kNN-TL/fairseq/dataclass/constants.py\n\tnew file: kNN-TL/fairseq/dataclass/initialize.py\n\tnew file: kNN-TL/fairseq/dataclass/utils.py\n\tnew file: kNN-TL/fairseq/distributed/__init__.py\n\tnew file: kNN-TL/fairseq/distributed/distributed_timeout_wrapper.py\n\tnew file: kNN-TL/fairseq/distributed/fully_sharded_data_parallel.py\n\tnew file: kNN-TL/fairseq/distributed/legacy_distributed_data_parallel.py\n\tnew file: kNN-TL/fairseq/distributed/module_proxy_wrapper.py\n\tnew file: kNN-TL/fairseq/distributed/tpu_distributed_data_parallel.py\n\tnew file: kNN-TL/fairseq/distributed/utils.py\n\tnew file: kNN-TL/fairseq/file_chunker_utils.py\n\tnew file: kNN-TL/fairseq/file_io.py\n\tnew file: kNN-TL/fairseq/file_utils.py\n\tnew file: kNN-TL/fairseq/hub_utils.py\n\tnew file: kNN-TL/fairseq/incremental_decoding_utils.py\n\tnew file: kNN-TL/fairseq/iterative_refinement_generator.py\n\tnew file: kNN-TL/fairseq/logging/__init__.py\n\tnew file: kNN-TL/fairseq/logging/meters.py\n\tnew file: kNN-TL/fairseq/logging/metrics.py\n\tnew file: kNN-TL/fairseq/logging/progress_bar.py\n\tnew file: kNN-TL/fairseq/model_parallel/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/criterions/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/criterions/vocab_parallel_cross_entropy.py\n\tnew file: kNN-TL/fairseq/model_parallel/megatron_trainer.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/pipeline_parallel_transformer/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/pipeline_parallel_transformer/layers.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/pipeline_parallel_transformer/model.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/roberta/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/roberta/model.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/transformer.py\n\tnew file: kNN-TL/fairseq/model_parallel/models/transformer_lm.py\n\tnew file: kNN-TL/fairseq/model_parallel/modules/__init__.py\n\tnew file: kNN-TL/fairseq/model_parallel/modules/multihead_attention.py\n\tnew file: kNN-TL/fairseq/model_parallel/modules/transformer_layer.py\n\tnew file: kNN-TL/fairseq/models/__init__.py\n\tnew file: kNN-TL/fairseq/models/bart/__init__.py\n\tnew file: kNN-TL/fairseq/models/bart/hub_interface.py\n\tnew file: kNN-TL/fairseq/models/bart/model.py\n\tnew file: kNN-TL/fairseq/models/composite_encoder.py\n\tnew file: kNN-TL/fairseq/models/distributed_fairseq_model.py\n\tnew file: kNN-TL/fairseq/models/ema/__init__.py\n\tnew file: kNN-TL/fairseq/models/ema/ema.py\n\tnew file: kNN-TL/fairseq/models/fairseq_decoder.py\n\tnew file: kNN-TL/fairseq/models/fairseq_encoder.py\n\tnew file: kNN-TL/fairseq/models/fairseq_incremental_decoder.py\n\tnew file: kNN-TL/fairseq/models/fairseq_model.py\n\tnew file: kNN-TL/fairseq/models/fconv.py\n\tnew file: kNN-TL/fairseq/models/fconv_lm.py\n\tnew file: kNN-TL/fairseq/models/fconv_self_att.py\n\tnew file: kNN-TL/fairseq/models/hubert/__init__.py\n\tnew file: kNN-TL/fairseq/models/hubert/hubert.py\n\tnew file: kNN-TL/fairseq/models/hubert/hubert_asr.py\n\tnew file: kNN-TL/fairseq/models/huggingface/__init__.py\n\tnew file: kNN-TL/fairseq/models/huggingface/hf_gpt2.py\n\tnew file: kNN-TL/fairseq/models/lightconv.py\n\tnew file: kNN-TL/fairseq/models/lightconv_lm.py\n\tnew file: kNN-TL/fairseq/models/lstm.py\n\tnew file: kNN-TL/fairseq/models/lstm_lm.py\n\tnew file: kNN-TL/fairseq/models/masked_lm.py\n\tnew file: kNN-TL/fairseq/models/model_utils.py\n\tnew file: kNN-TL/fairseq/models/multilingual_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/__init__.py\n\tnew file: kNN-TL/fairseq/models/nat/cmlm_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/fairseq_nat_model.py\n\tnew file: kNN-TL/fairseq/models/nat/insertion_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/iterative_nonautoregressive_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/levenshtein_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/levenshtein_utils.py\n\tnew file: kNN-TL/fairseq/models/nat/nat_crf_transformer.py\n\tnew file: kNN-TL/fairseq/models/nat/nonautoregressive_ensembles.py\n\tnew file: kNN-TL/fairseq/models/na…","shortMessageHtmlLink":"\tmodified: .DS_Store"}},{"before":null,"after":"913333ddb1d1ca7c41dc007d6af3d295f57f191a","ref":"refs/heads/main","pushedAt":"2023-07-04T07:40:41.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"sudanl","name":"Shudong Liu","path":"/sudanl","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57032474?s=80&v=4"},"commit":{"message":"\tmodified: .DS_Store\n\tmodified: README.md","shortMessageHtmlLink":"\tmodified: .DS_Store"}}],"hasNextPage":false,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAADXg__zQA","startCursor":null,"endCursor":null}},"title":"Activity · NLP2CT/kNN-TL"}