Menjalankan inferensi LLM di GPU Cloud Run dengan Hugging Face TGI (layanan)
Tetap teratur dengan koleksi
Simpan dan kategorikan konten berdasarkan preferensi Anda.
Contoh berikut menunjukkan cara menjalankan layanan backend yang menjalankan toolkit Hugging Face Text Generation Inference (TGI), yang merupakan toolkit untuk men-deploy dan menayangkan Large Language Model (LLM), menggunakan Llama 3.
[[["Mudah dipahami","easyToUnderstand","thumb-up"],["Memecahkan masalah saya","solvedMyProblem","thumb-up"],["Lainnya","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Masalah terjemahan","translationIssue","thumb-down"],["Lainnya","otherDown","thumb-down"]],["Terakhir diperbarui pada 2024-12-21 UTC."],[],[]]