Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add alpha scaling to lora #8248

Merged
merged 13 commits into from
Feb 25, 2024
Merged

add alpha scaling to lora #8248

merged 13 commits into from
Feb 25, 2024

Commits on Jan 17, 2024

  1. removed pdeprecated eft model

    Signed-off-by: arendu <adithya.r@gmail.com>
    arendu committed Jan 17, 2024
    Configuration menu
    Copy the full SHA
    00a164c View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    fe953a9 View commit details
    Browse the repository at this point in the history

Commits on Jan 21, 2024

  1. Configuration menu
    Copy the full SHA
    e369f3b View commit details
    Browse the repository at this point in the history

Commits on Jan 22, 2024

  1. Configuration menu
    Copy the full SHA
    881fcdc View commit details
    Browse the repository at this point in the history

Commits on Jan 25, 2024

  1. Configuration menu
    Copy the full SHA
    e9b9a15 View commit details
    Browse the repository at this point in the history
  2. add alpha

    Signed-off-by: arendu <adithya.r@gmail.com>
    arendu committed Jan 25, 2024
    Configuration menu
    Copy the full SHA
    7361e06 View commit details
    Browse the repository at this point in the history
  3. default for alpha

    Signed-off-by: arendu <adithya.r@gmail.com>
    arendu committed Jan 25, 2024
    Configuration menu
    Copy the full SHA
    c6a9fa1 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    9ee0566 View commit details
    Browse the repository at this point in the history

Commits on Feb 14, 2024

  1. Configuration menu
    Copy the full SHA
    37d79d5 View commit details
    Browse the repository at this point in the history

Commits on Feb 23, 2024

  1. add alpha scaling to lora (#8483)

    * coldfix (#8412)
    
    Signed-off-by: George Zelenfroynd <gzelenfroind@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Fixed errors in the CTM gen functions (#8416) (#8420)
    
    Signed-off-by: Taejin Park <tango4j@gmail.com>
    Co-authored-by: Taejin Park <tango4j@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add change_vocabulary and save_tokenizers() support to Multitask ASR models (#8357) (#8367)
    
    * Add change_vocabulary and save_tokenizers() support
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * Update nemo/collections/asr/models/aed_multitask_models.py
    
    ---------
    
    Signed-off-by: smajumdar <titu1994@gmail.com>
    Signed-off-by: Somshubra Majumdar <titu1994@gmail.com>
    Co-authored-by: Somshubra Majumdar <titu1994@gmail.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Co-authored-by: Piotr Żelasko <petezor@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * fix path location and branch (#8314)
    
    * fix path location and branch (#8304)
    
    * fix path location and branch
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    
    * change to a floating point number
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    
    ---------
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    Co-authored-by: Nithin Rao Koluguri <nithinraok>
    Co-authored-by: Somshubra Majumdar <titu1994@gmail.com>
    
    * updat ebranch in tutorial
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    
    ---------
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    Co-authored-by: Nithin Rao <nithinrao.koluguri@gmail.com>
    Co-authored-by: Somshubra Majumdar <titu1994@gmail.com>
    Co-authored-by: Nithin Rao Koluguri <nithinraok>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add TP comm overlap knobs to AutocastTransformerLayer (#8290)
    
    Signed-off-by: Jaemin Choi <jaeminc@nvidia.com>
    Co-authored-by: Jaemin Choi <jaeminc@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * add deallocate pipeline output optimization (#8279) (#8318)
    
    * add deallocate pipeline output optimization
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    ---------
    
    Signed-off-by: Jimmy Zhang <jiemingz@nvidia.com>
    Co-authored-by: JimmyZhang12 <67203904+JimmyZhang12@users.noreply.github.com>
    Co-authored-by: Jimmy Zhang <jiemingz@nvidia.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * remove assertion (#8302) (#8321)
    
    Signed-off-by: dimapihtar <dpihtar@gmail.com>
    Co-authored-by: Dmytro Pykhtar <37850217+dimapihtar@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Keep max_seqlen and cu_seqlens_argmin for later micro-batches when PP>1 (#8334) (#8346)
    
    Signed-off-by: Sangkug Lym <slym@nvidia.com>
    Co-authored-by: Sangkug Lym <slym@nvidia.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Enable megatron core loggers for GPT pretraining (#8354) (#8384)
    
    * Logging changes tested for gpt_pretraining
    
    * Additional args
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    ---------
    
    Signed-off-by: Aishwarya Bhandare <abhandare@nvidia.com>
    Co-authored-by: ashbhandare <ash.bhandare@gmail.com>
    Co-authored-by: Aishwarya Bhandare <abhandare@nvidia.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Fix dreambooth data sampler issue (#8400) (#8413)
    
    * Turn on drop last
    
    * Some neva fixes
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    ---------
    
    Signed-off-by: yaoyu-33 <yaoyu.094@gmail.com>
    Co-authored-by: yaoyu-33 <54727607+yaoyu-33@users.noreply.github.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * add ensemble decoding fix (#8427) (#8433)
    
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    Co-authored-by: Nithin Rao <nithinrao.koluguri@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * NeVA Tutorial Notebook (#8217)
    
    * init commit - neva tutorial
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * NeVA tutorial notebook
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * init commit - neva tutorial
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Signed-off-by: Pratyush Muthukumar <pannumuthu@gmail.com>
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * NeVA tutorial notebook
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Signed-off-by: Pratyush Muthukumar <pannumuthu@gmail.com>
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * requested changes
    
    Signed-off-by: Pratyush Muthukumar <pannumuthu@gmail.com>
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * add inference via script
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * requested changes
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * requested changes
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    * add codeblocks to run torchrun in notebook
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    
    ---------
    
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Signed-off-by: Pratyush Muthukumar <pannumuthu@gmail.com>
    Co-authored-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * mcore customization doc minor fix (#8421) (#8437)
    
    Signed-off-by: Huiying Li <willwin.lee@gmail.com>
    Co-authored-by: Huiying <willwin.lee@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add `loop_labels` algorithm for TDT greedy decoding (#8215)
    
    * Add `loop_labels` algorithm for TDT greedy decoding
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Use `loop_labels` by default
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Loop labels greedy decoding v2
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add comments. Clean up
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add comments
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add comments
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add tests for batched hypotheses
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add tests for batched alignments
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add comments
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Fix comment
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Fix test
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add computer for TDT
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Fix TDT decoding algorithm
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Use loop frames by default for TDT
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Remove "loop frames" implementation for TDT
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Clean up
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Add comments
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    * Fix confidence. Use tensor for durations.
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    
    ---------
    
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add dist ckpt support for regular optimizers (#7749) (#8293)
    
    * Add dist ckpt support for regular optimizers
    
    * [tutorial] fixed missing RIR scripts file. (#8257)
    
    * fix imports
    
    * imports fix
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * ci imports fix
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * revert asr notebook
    
    * revert asr notebook
    
    ---------
    
    Signed-off-by: Mikołaj Błaż <mblaz@nvidia.com>
    Signed-off-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Signed-off-by: dimapihtar <dpihtar@gmail.com>
    Co-authored-by: mikolajblaz <mikolajblaz@users.noreply.github.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Co-authored-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Co-authored-by: Dmytro Pykhtar <37850217+dimapihtar@users.noreply.github.com>
    Co-authored-by: dimapihtar <dpihtar@gmail.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Multimodal r1.23.0 bug fix  (#8315) (#8339)
    
    * Rename quick-gelu
    
    * ddpm config guard
    
    * Fix ddpm edit api
    
    * Fix insert_image_token cfg issue
    
    * neva updates
    
    * reformat
    
    * Add back jenkins
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * Fix jenkins
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * Fix bugs
    
    * Update default neva template
    
    ---------
    
    Signed-off-by: yaoyu-33 <yaoyu.094@gmail.com>
    Co-authored-by: yaoyu-33 <54727607+yaoyu-33@users.noreply.github.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * mcore ds fix (#8283) (#8385)
    
    * [tutorial] fixed missing RIR scripts file. (#8257)
    
    * add values to en tts dict (#7879)
    
    * mcore ds fix
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * update mcore
    
    * revert asr files
    
    * add comments
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * add support for mcore mock dataset
    
    * update mcore version
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * update gpt cfg
    
    * update mcore commit
    
    * fix Bert unit tests
    
    * update bert tests
    
    * fix bert mcore test
    
    * fix gpt jenkins tests
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * update apex & TE commits
    
    * revert apex installation
    
    * turn off the fusion for jenkins
    
    ---------
    
    Signed-off-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Signed-off-by: Mariana Graterol Fuenmayor <marianag@nvidia.com>
    Signed-off-by: Dmytro Pykhtar <dpykhtar@login-eos01.eos.clusters.nvidia.com>
    Signed-off-by: dimapihtar <dpihtar@gmail.com>
    Co-authored-by: Dmytro Pykhtar <37850217+dimapihtar@users.noreply.github.com>
    Co-authored-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Co-authored-by: Mariana <47233618+mgrafu@users.noreply.github.com>
    Co-authored-by: Dmytro Pykhtar <dpykhtar@login-eos01.eos.clusters.nvidia.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Co-authored-by: Pablo Garay <palenq@gmail.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * MCore dataset compatibility for tokenizers (#8390) (#8397)
    
    * Add unique_identifiers for all tokenizers and eod for SentencePieceTokenizer
    
    * Add generalized token aliases to TokenizerSpec to conform with MegatronTokenizer's interface. Remove now-redundant individual fixes from AutoTokenizer and SentencePieceTokenizer.
    
    ---------
    
    Signed-off-by: Valerie Sarge <vsarge@nvidia.com>
    Co-authored-by: Valerie Sarge <vsarge@nvidia.com>
    Co-authored-by: Pablo Garay <palenq@gmail.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Canary: inference tokenization improvements; preserving custom keys when creating tarred manifests (#8432)
    
    * Improvements for Canary:
    
    - carry over custom keys when creatin tarred manifests
    - selectable text field in ASR eval
    - get rid of prompt slicing, create proper inference prompts
    
    Signed-off-by: Piotr Żelasko <petezor@gmail.com>
    
    * set ensure_ascii=False in tarred conversion to avoid breaking tokenizers trained on UTF-8 encoding
    
    Signed-off-by: Piotr Żelasko <petezor@gmail.com>
    
    ---------
    
    Signed-off-by: Piotr Żelasko <petezor@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * add  sbert to IR (#8445)
    
    * add  sbert to IR
    
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    
    * add doc
    
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    
    * fix the  auto_tokenizer property method reset bug
    
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    
    * addressed bot comments
    
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    ---------
    
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Update readme (#8440)
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * udpate
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * landing pages added
    
    * landing page added for vision
    
    * landing pages updated
    
    * some minor changes to the main readme
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    * typo fixed
    
    * update
    
    Signed-off-by: eharper <eharper@nvidia.com>
    
    ---------
    
    Signed-off-by: eharper <eharper@nvidia.com>
    Co-authored-by: ntajbakhsh <ntajbakhsh@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * NeMo-Mistral to HF converter bugfix. (#8353) (#8442)
    
    Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
    Co-authored-by: akoumpa <153118171+akoumpa@users.noreply.github.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Fixing mcore bert for TP, PP and SP (#8336) (#8443)
    
    * Fixing mcore bert for TP, PP and SP
    
    * Fixing mcore bert for TP, PP and SP
    
    * Fixing mcore version
    
    * Fixing mcore version
    
    * Update Jenkinsfile
    
    * Update Jenkinsfile
    
    * Update Jenkinsfile
    
    ---------
    
    Signed-off-by: Shanmugam Ramasamy <111910568+shanmugamr1992@users.noreply.github.com>
    Co-authored-by: Shanmugam Ramasamy <111910568+shanmugamr1992@users.noreply.github.com>
    Co-authored-by: Shanmugam Ramasamy <shanmugamr@shanmugamr-mlt.client.nvidia.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add LoRA support to all linear layers (#7988)
    
    * Added LoRA support for the Dense layer of Attention
    
    * Added LoRA MLP support to MCore and NeMo models.
    
    * Change LoRA config default to QKV.
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * Fixed bug with ddp training.
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * MCoreMixin chages.
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * using new commit of meg-LM
    
    Signed-off-by: arendu <adithya.r@gmail.com>
    
    * add cpu_offloading_num_layers to conversion script until bug in megatron is fixed
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    * fix peft mixin arguments to follow mcore 0.5
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    * [pre-commit.ci] auto fixes from pre-commit.com hooks
    
    for more information, see https://pre-commit.ci
    
    * update megatron commit to fix ci error
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    * try to fix ci
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    * try to fix ci
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    * add cfg default
    
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    
    ---------
    
    Signed-off-by: Adi Renduchintala <adithyare@nvidia.com>
    Signed-off-by: Jiaqi Zeng <jiaqiz@nvidia.com>
    Signed-off-by: arendu <adithya.r@gmail.com>
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Co-authored-by: Adi Renduchintala <adithyare@nvidia.com>
    Co-authored-by: Jiaqi Zeng <jiaqiz@nvidia.com>
    Co-authored-by: arendu <adithya.r@gmail.com>
    Co-authored-by: HeyyyyyyG <49757268+HeyyyyyyG@users.noreply.github.com>
    Co-authored-by: Chen Cui <chcui@nvidia.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Add Neva Template for NV-DPO Models  (#8358)
    
    * add/rename from nvgpt to nv_steerlm, add nv_dpo template
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    
    * add nv_dpo conversation to accomendate empty system message
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    
    * handle nv_dpo template text generation
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    
    * add prompt string to nvgpt
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    
    * bugfix for inference prompt template
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    
    * bug fix for grabbing clean text
    
    Signed-off-by: Huiying Li <willwin.lee@gmail.com>
    
    * fix code format
    
    Signed-off-by: Huiying Li <willwin.lee@gmail.com>
    
    ---------
    
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    Signed-off-by: Huiying Li <willwin.lee@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Rebase scaling alpha
    
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * default for alpha
    
    Signed-off-by: arendu <adithya.r@gmail.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    * Rebase scaling alpha
    
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    
    ---------
    
    Signed-off-by: George Zelenfroynd <gzelenfroind@nvidia.com>
    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    Signed-off-by: Taejin Park <tango4j@gmail.com>
    Signed-off-by: smajumdar <titu1994@gmail.com>
    Signed-off-by: Somshubra Majumdar <titu1994@gmail.com>
    Signed-off-by: Nithin Rao Koluguri <nithinraok>
    Signed-off-by: Jaemin Choi <jaeminc@nvidia.com>
    Signed-off-by: Jimmy Zhang <jiemingz@nvidia.com>
    Signed-off-by: dimapihtar <dpihtar@gmail.com>
    Signed-off-by: Sangkug Lym <slym@nvidia.com>
    Signed-off-by: Aishwarya Bhandare <abhandare@nvidia.com>
    Signed-off-by: yaoyu-33 <yaoyu.094@gmail.com>
    Signed-off-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Signed-off-by: Pratyush Muthukumar <pannumuthu@gmail.com>
    Signed-off-by: Huiying Li <willwin.lee@gmail.com>
    Signed-off-by: Vladimir Bataev <vbataev@nvidia.com>
    Signed-off-by: Mikołaj Błaż <mblaz@nvidia.com>
    Signed-off-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Signed-off-by: Mariana Graterol Fuenmayor <marianag@nvidia.com>
    Signed-off-by: Dmytro Pykhtar <dpykhtar@login-eos01.eos.clusters.nvidia.com>
    Signed-off-by: Valerie Sarge <vsarge@nvidia.com>
    Signed-off-by: Piotr Żelasko <petezor@gmail.com>
    Signed-off-by: ataghibakhsh <ataghibakhsh@nvidia.com>
    Signed-off-by: eharper <eharper@nvidia.com>
    Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
    Signed-off-by: Shanmugam Ramasamy <111910568+shanmugamr1992@users.noreply.github.com>
    Signed-off-by: Adi Renduchintala <adithyare@nvidia.com>
    Signed-off-by: Jiaqi Zeng <jiaqiz@nvidia.com>
    Signed-off-by: arendu <adithya.r@gmail.com>
    Signed-off-by: Chen Cui <chcui@nvidia.com>
    Signed-off-by: HuiyingLi <willwin.lee@gmail.com>
    Co-authored-by: George <37293288+Jorjeous@users.noreply.github.com>
    Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
    Co-authored-by: Taejin Park <tango4j@gmail.com>
    Co-authored-by: Somshubra Majumdar <titu1994@gmail.com>
    Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
    Co-authored-by: Piotr Żelasko <petezor@gmail.com>
    Co-authored-by: Nithin Rao <nithinrao.koluguri@gmail.com>
    Co-authored-by: Jaemin Choi <minitu77@gmail.com>
    Co-authored-by: Jaemin Choi <jaeminc@nvidia.com>
    Co-authored-by: JimmyZhang12 <67203904+JimmyZhang12@users.noreply.github.com>
    Co-authored-by: Jimmy Zhang <jiemingz@nvidia.com>
    Co-authored-by: Dmytro Pykhtar <37850217+dimapihtar@users.noreply.github.com>
    Co-authored-by: Sangkug Lym <slym@nvidia.com>
    Co-authored-by: Eric Harper <complex451@gmail.com>
    Co-authored-by: ashbhandare <ash.bhandare@gmail.com>
    Co-authored-by: Aishwarya Bhandare <abhandare@nvidia.com>
    Co-authored-by: yaoyu-33 <54727607+yaoyu-33@users.noreply.github.com>
    Co-authored-by: Pratyush Muthukumar <30813477+PannuMuthu@users.noreply.github.com>
    Co-authored-by: Pratyush Muthukumar <pmuthukumar@nvidia.com>
    Co-authored-by: Huiying <willwin.lee@gmail.com>
    Co-authored-by: Vladimir Bataev <vbataev@nvidia.com>
    Co-authored-by: mikolajblaz <mikolajblaz@users.noreply.github.com>
    Co-authored-by: Xuesong Yang <1646669+XuesongYang@users.noreply.github.com>
    Co-authored-by: dimapihtar <dpihtar@gmail.com>
    Co-authored-by: Mariana <47233618+mgrafu@users.noreply.github.com>
    Co-authored-by: Dmytro Pykhtar <dpykhtar@login-eos01.eos.clusters.nvidia.com>
    Co-authored-by: Pablo Garay <palenq@gmail.com>
    Co-authored-by: Valerie Sarge <vsarge@nvidia.com>
    Co-authored-by: Ali Taghibakhshi <71892896+JRD971000@users.noreply.github.com>
    Co-authored-by: ntajbakhsh <ntajbakhsh@nvidia.com>
    Co-authored-by: akoumpa <153118171+akoumpa@users.noreply.github.com>
    Co-authored-by: Shanmugam Ramasamy <111910568+shanmugamr1992@users.noreply.github.com>
    Co-authored-by: Shanmugam Ramasamy <shanmugamr@shanmugamr-mlt.client.nvidia.com>
    Co-authored-by: Tugrul Konuk <ertkonuk@gmail.com>
    Co-authored-by: Adi Renduchintala <adithyare@nvidia.com>
    Co-authored-by: Jiaqi Zeng <jiaqiz@nvidia.com>
    Co-authored-by: arendu <adithya.r@gmail.com>
    Co-authored-by: HeyyyyyyG <49757268+HeyyyyyyG@users.noreply.github.com>
    Co-authored-by: Chen Cui <chcui@nvidia.com>
    Configuration menu
    Copy the full SHA
    f655aaa View commit details
    Browse the repository at this point in the history
  2. Merge branch 'main' into adithyare/lora_alpha

    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    michal2409 authored Feb 23, 2024
    Configuration menu
    Copy the full SHA
    7db6e06 View commit details
    Browse the repository at this point in the history

Commits on Feb 24, 2024

  1. Merge branch 'main' into adithyare/lora_alpha

    Signed-off-by: Michal Futrega <mfutrega@nvidia.com>
    michal2409 authored Feb 24, 2024
    Configuration menu
    Copy the full SHA
    6dc3b51 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    67dac15 View commit details
    Browse the repository at this point in the history