github.com/NVIDIA/TensorRT-LLM

Preview meta tags from the github.com website.

Linked Hostnames

28

Thumbnail

Search Engine Appearance

Google

https://github.com/NVIDIA/TensorRT-LLM

GitHub - NVIDIA/TensorRT-LLM: TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way. - NVIDIA/TensorRT-LLM



Bing

GitHub - NVIDIA/TensorRT-LLM: TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

https://github.com/NVIDIA/TensorRT-LLM

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way. - NVIDIA/TensorRT-LLM



DuckDuckGo

https://github.com/NVIDIA/TensorRT-LLM

GitHub - NVIDIA/TensorRT-LLM: TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way. - NVIDIA/TensorRT-LLM

  • General Meta Tags

    46
    • title
      GitHub - NVIDIA/TensorRT-LLM: TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.
    • charset
      utf-8
    • route-pattern
      /:user_id/:repository
    • route-controller
      files
    • route-action
      disambiguate
  • Open Graph Meta Tags

    9
    • og:image
      https://opengraph.githubassets.com/1cc6792cea9a808dbd43fc1308006595d21bb93b7642b6a034507a08a3446e26/NVIDIA/TensorRT-LLM
    • og:image:alt
      TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorR...
    • og:image:width
      1200
    • og:image:height
      600
    • og:site_name
      GitHub
  • Twitter Meta Tags

    5
    • twitter:image
      https://opengraph.githubassets.com/1cc6792cea9a808dbd43fc1308006595d21bb93b7642b6a034507a08a3446e26/NVIDIA/TensorRT-LLM
    • twitter:site
      @github
    • twitter:card
      summary_large_image
    • twitter:title
      GitHub - NVIDIA/TensorRT-LLM: TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.
    • twitter:description
      TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorR...
  • Link Tags

    47
    • alternate icon
      https://github.githubassets.com/favicons/favicon.png
    • assets
      https://github.githubassets.com/
    • canonical
      https://github.com/NVIDIA/TensorRT-LLM
    • dns-prefetch
      https://github.githubassets.com
    • dns-prefetch
      https://avatars.githubusercontent.com

Links

231