docs.inferless.com/how-to-guides/deploy-a-codellama-python-34b-model-using-inferless

Preview meta tags from the docs.inferless.com website.

Linked Hostnames

7

Thumbnail

Search Engine Appearance

Google

https://docs.inferless.com/how-to-guides/deploy-a-codellama-python-34b-model-using-inferless

Deploy a CodeLlama-Python-34B Model using Inferless - Inferless

In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.



Bing

Deploy a CodeLlama-Python-34B Model using Inferless - Inferless

https://docs.inferless.com/how-to-guides/deploy-a-codellama-python-34b-model-using-inferless

In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.



DuckDuckGo

https://docs.inferless.com/how-to-guides/deploy-a-codellama-python-34b-model-using-inferless

Deploy a CodeLlama-Python-34B Model using Inferless - Inferless

In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.

  • General Meta Tags

    15
    • title
      Deploy a CodeLlama-Python-34B Model using Inferless - Inferless
    • charset
      utf-8
    • viewport
      width=device-width, initial-scale=1
    • next-size-adjust
    • description
      In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.
  • Open Graph Meta Tags

    6
    • og:title
      Deploy a CodeLlama-Python-34B Model using Inferless - Inferless
    • og:description
      In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.
    • og:image
      https://inferless-68.mintlify.app/mintlify-assets/_next/image?url=%2Fapi%2Fog%3Fdivision%3DHow-to%2BGuides%26title%3DDeploy%2Ba%2BCodeLlama-Python-34B%2BModel%2Busing%2BInferless%26description%3DIn%2Bthis%2Btutorial%252C%2Bwe%2527ll%2Bshow%2Bthe%2Bdeployment%2Bprocess%2Bof%2Ba%2Bquantized%2BGPTQ%2Bmodel%2Busing%2BvLLM.%2BWe%2Bare%2Bdeploying%2Ba%2BGPTQ%252C%2B4-bit%2Bquantized%2Bversion%2Bof%2Bthe%2BcodeLlama-Python-34B%2Bmodel.%26logoLight%3Dhttps%253A%252F%252Fmintlify.s3.us-west-1.amazonaws.com%252Finferless-68%252Flogo.svg%26logoDark%3Dhttps%253A%252F%252Fmintlify.s3.us-west-1.amazonaws.com%252Finferless-68%252Flogo.svg%26primaryColor%3D%252394CF09%26lightColor%3D%25239FDD0C%26darkColor%3D%252394CF09%26backgroundLight%3D%2523ffffff%26backgroundDark%3D%25230c0d0b&w=1200&q=100
    • og:image:width
      1200
    • og:image:height
      630
  • Twitter Meta Tags

    8
    • twitter:card
      summary_large_image
    • twitter:title
      Deploy a CodeLlama-Python-34B Model using Inferless - Inferless
    • twitter:card
      summary_large_image
    • twitter:title
      Deploy a CodeLlama-Python-34B Model using Inferless - Inferless
    • twitter:description
      In this tutorial, we'll show the deployment process of a quantized GPTQ model using vLLM. We are deploying a GPTQ, 4-bit quantized version of the codeLlama-Python-34B model.
  • Link Tags

    18
    • alternate
      /sitemap.xml
    • apple-touch-icon
      https://mintlify.s3-us-west-1.amazonaws.com/inferless-68/_generated/favicon/apple-touch-icon.png?v=3
    • icon
      https://mintlify.s3-us-west-1.amazonaws.com/inferless-68/_generated/favicon/favicon-32x32.png?v=3
    • icon
      https://mintlify.s3-us-west-1.amazonaws.com/inferless-68/_generated/favicon/favicon-16x16.png?v=3
    • preload
      /mintlify-assets/_next/static/media/bb3ef058b751a6ad-s.p.woff2

Links

67