Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sync master #1993

Merged
merged 24 commits into from
Dec 20, 2024
Merged

Sync master #1993

merged 24 commits into from
Dec 20, 2024

Conversation

arjunsuresh
Copy link
Contributor

No description provided.

nvzhihanj and others added 24 commits December 9, 2024 14:32
* Fix loadgen build for version numbers having "0"

* Update test-resnet50.yml

* Update test-retinanet.yml

* Update test-bert.yml
* Fix submission checker for v5.0 rgat

* Update submission_checker.py | Updates for v5.0

* [Automated Commit] Format Codebase

* Update submission_checker.py | Fixes latency_constraints for v5.0

* [Automated Commit] Format Codebase

---------

Co-authored-by: mlcommons-bot <[email protected]>
* Fix llama3-405B docker workflow

* Fix the performance sample count from 8312 to 8313

* More fixes
* Fix submission checker for v5.0 rgat

* Fix accuracy pattern for rgat, report-generator for v5.0
* More minor fixes

* Fix indentation for stats report
* Require equal issue mode for R-GAT

* Add equal issue note in readme

---------

Co-authored-by: Miro <[email protected]>
* Fixes #1648, restrict loadgen uncommitted error message to within the loadgen directory

* Update test-rnnt.yml (#1688)

Stopping the github action for rnnt

* Added docs init

Added github action for website publish

Update benchmark documentation

Update publish.yaml

Update publish.yaml

Update benchmark documentation

Improved the submission documentation

Fix taskname

Removed unused images

* Fix benchmark URLs

* Fix links

* Add _full variation to run commands

* Added script flow diagram

* Added docker setup command for CM, extra run options

* Added support for docker options in the docs

* Added --quiet to the CM run_cmds in docs

* Fix the test query count for cm commands

* Support ctuning-cpp implementation

* Added commands for mobilenet models

* Docs cleanup

* Docs cleanup

* Added separate files for dataset and models in the docs

* Remove redundant tab in the docs

* Fixes some WIP models in the docs

* Use the official docs page for CM installation

* Fix the deadlink in docs

* Fix indendation issue in docs

* Added dockerinfo for nvidia implementation

* Added run options for gptj

* Added execution environment tabs

* Cleanup of the docs

* Cleanup of the docs

* Reordered the sections of the docs page

* Removed an unnecessary heading in the docs

* Fixes the commands for datacenter

* Fix the build --sdist for loadgen

* Fixes #1761, llama2 and mixtral runtime error on CPU systems

* Added mixtral to the benchmark list, improved benchmark docs

* Update docs for MLPerf inference v4.1

* Update docs for MLPerf inference v4.1

* Fix typo

* Gave direct link to implementation readmes

* Added tables detailing implementations

* Update vision README.md, split the frameworks into separate rows

* Update README.md

* pointed links to specific frameworks

* pointed links to specific frameworks

* Update Submission_Guidelines.md

* Update Submission_Guidelines.md

* Update Submission_Guidelines.md

* api support llama2

* Added request module and reduced max token len

* Fix for llama2 api server

* Update SUT_API offline to work for OpenAI

* Update SUT_API.py

* Minor fixes

* Fix json import in SUT_API.py

* Fix llama2 token length

* Added model name verification with server

* clean temp files

* support num_workers in LLAMA2 SUTs

* Remove batching from Offline SUT_API.py

* Update SUT_API.py

* Minor fixes for llama2 API

* Fix for llama2 API

* removed table of contents

* enabled llama2-nvidia + vllm-NM : WIP

* enabled dlrm for intel

* lower cased implementation

* added raw data input

* corrected data download commands

* renamed filename

* changes for bert and vllm

* documentation to work on custom repo and branch

* benchmark index page update

* enabled sdxl for nvidia and intel

* updated vllm server run cmd

* benchmark page information addition

* fix indendation issue

* Added submission categories

* update submission page - generate submission with or w/o using CM for benchmarking

* Updated kits dataset documentation

* Updated model parameters

* updation of information

* updated non cm based benchmark

* added info about hf password

* added links to model and access tokens

* Updated reference results structuree tree

* submission docs cleanup

* Some cleanups for benchmark info

* Some cleanups for benchmark info

* Some cleanups for benchmark info

* added generic stubs deepsparse

* Some cleanups for benchmark info

* Some cleanups for benchmark info

* Some cleanups for benchmark info

* Some cleanups for benchmark info (FID and CLIP data added)

* typo fix for bert deepsparse framework

* added min system requirements for models

* fixed code version

* changes for displaying reference and intel implementation tip

* added reference to installation page

* updated neural magic documentation

* Added links to the install page, redirect benchmarks page

* added tips about batch size and dataset for nvidia llama2

* fix conditions logic

* modified tips and additional run cmds

* sentence corrections

* Minor fix for the documentation

* fixed bug in deepsparse generic model stubs + styling

* added more information to stubs

* Added SCC24 readme, support reproducibility in the docs

* Made clear the custom CM repo URL format

* Support conditional implementation, setup and run tips

* Support rocm for sdxl

* Fix _short tag support

* Fix install URL

* Expose bfloat16 and float16 options for sdxl

* Expose download model to host option for sdxl

* IndySCC24 documentation added

* Improve the SCC24 docs

* Improve the support of short variation

* Improved the indyscc24 documentation

* Updated scc run commands

* removed test_query_count option for scc

* Remove scc24 in the main docs

* Remove scc24 in the main docs

* Fix docs: indendation issue on the submission page

* generalised code for skipping test query count

* Fixes for SCC24 docs

* Fix scenario text in main.py

* Fix links for scc24

* Fix links for scc24

* Improve the general docs

* Fix links for scc24

* Use float16 in scc24 doc

* Improve scc24 docs

* Improve scc24 docs

* Use float16 in scc24 doc

* fixed command bug

* Fix typo in docs

* Fix typo in docs

* Remove unnecessary indendation in docs

* initial commit for tip - native run CUDA

* Updated tip

* added docker_cm_repo_branch to more run option - docker

* Update docs for IndySCC24

* Support custom repo branch and owner for final report generation

* enabled amd implementation for llama2

* updations for amd - docs

* Fix scenarios in docs page

* formatted the files to pass the gh action

* scenarios -> fixed_scenarios in docs

* [Automated Commit] Format Codebase

* Update indyscc24-bert.md

* Update scc24.md

* updated tip for reference implementation (#1912)

* [Automated Commit] Format Codebase

* fix for run suffix (#1913)

* [Automated Commit] Format Codebase

* Updation for adding submission flow diagram

* Added submission flow diagram

* Update scc24.md

* changes in submission documentation (#1946)

* update results category (#1947)

* changes for adding rgat to docs (#1965)

* Update index.md | Added R-GAT details (WIP)

* Update index.md

* Create system_requirements.yml

* Update system_requirements.yml

* Update system_requirements.yml

* Update system_requirements.yml

---------

Co-authored-by: anandhu-eng <[email protected]>
Co-authored-by: ANANDHU S <[email protected]>
Co-authored-by: Michael Goin <[email protected]>
Co-authored-by: arjunsuresh <[email protected]>
Co-authored-by: Pablo Gonzalez <[email protected]>
Co-authored-by: Mitchelle Rasquinha <[email protected]>
Co-authored-by: Miro <[email protected]>
* Update automated run command section

* add cm commands for model and dataset downloads

* Update README.md

* Update cm run cmds

---------

Co-authored-by: Miro <[email protected]>
* Unify llama3 names to llama3.1-405b

* Set mlperf.conf name to llama3_1-405b
* Create test-rgat.yml

* Update test-rgat.yml

* Update test-rgat.yml

---------

Co-authored-by: Miro <[email protected]>
* Create benchmark-checklist.md for r-gat

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

* Update benchmark-checklist.md

---------

Co-authored-by: Miro <[email protected]>
@arjunsuresh arjunsuresh requested a review from a team as a code owner December 20, 2024 19:09
Copy link

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@arjunsuresh arjunsuresh merged commit 9309ef7 into dev Dec 20, 2024
34 of 35 checks passed
@github-actions github-actions bot locked and limited conversation to collaborators Dec 20, 2024
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants