|
Python |
5 |
Inference code for LLaMA models |
Mar 20, 2023 |
|
Python |
4 |
Inference code for LLaMA models |
Apr 18, 2023 |
|
Python |
2 |
Inference code for LLaMA models |
Apr 25, 2023 |
|
None |
4 |
Inference code for LLaMA models |
Apr 18, 2023 |
|
Python |
17218 |
Inference code for LLaMA models |
Apr 10, 2023 |
|
Python |
4 |
Inference code for LLaMA models |
Mar 21, 2023 |
|
Python |
2 |
Inference code for LLaMA models |
Apr 19, 2023 |
|
Jupyter Notebook |
26 |
Inference code for LLaMA models |
Apr 22, 2023 |
|
Python |
174 |
Inference code for LLaMA models |
Apr 24, 2023 |
|
Python |
2 |
Inference code for LLaMA models |
Mar 09, 2023 |
|
Python |
3 |
Inference code for LLaMA models |
Mar 21, 2023 |
|
None |
2 |
Inference code for LLaMA models |
Jul 05, 2023 |
|
Python |
3 |
Inference code for LLaMA models |
Jul 21, 2023 |
|
None |
2 |
Inference code for LLaMA models |
May 24, 2023 |
|
Python |
962 |
Quantized inference code for LLaMA models |
Apr 23, 2023 |
|
Python |
3 |
Quantized inference code for LLaMA models |
May 13, 2023 |
|
Python |
5 |
[NOT TESTED!] Inference code for LLaMA models |
Mar 16, 2023 |
|
Python |
93 |
Inference on CPU code for LLaMA models |
Apr 27, 2023 |
|
Python |
60 |
Inference code for LLaMA models in JAX |
May 08, 2023 |
|
Python |
11 |
Inference code for LLaMA models (modified for cpu) |
Apr 29, 2023 |
|
Python |
102 |
Inference code for facebook LLaMA models with Wrapyfi support |
May 01, 2023 |
|
Python |
89 |
Inference script for Meta's LLaMA models using Hugging Face wrapper |
Apr 05, 2023 |
|
Python |
18 |
llama inference for tencentpretrain |
Apr 25, 2023 |
|
Python |
2 |
LLaMa model inference |
Mar 08, 2023 |
|
Rust |
315 |
Rust+OpenCL+AVX2 implementation of LLaMA inference code |
Apr 10, 2023 |
|
Python |
25 |
Tiny inference-only implementation of LLaMA |
Mar 16, 2023 |
|
Rust |
40 |
Using Code Language Models for Gradual Type Inference |
Apr 10, 2023 |
|
Python |
2 |
Tuning Llama Models > 7b |
Jun 19, 2023 |
|
None |
5 |
Run LLaMA inference on Apple Silicon GPUs. |
May 29, 2023 |
|
C++ |
2 |
Simple chat program for LLaMa models |
Jun 26, 2023 |
|
C |
3 |
Fast LLaMa inference on CPU using llama.cpp for Python |
Apr 07, 2023 |
|
Python |
18 |
Simple and fast server for GPTQ-quantized LLaMA inference |
May 24, 2023 |
|
Python |
10 |
Code for training & inference with FLAN family of models |
Apr 24, 2023 |
|
Python |
221 |
Guide for fine-tuning Llama/CodeLlama models |
Jan 18, 2024 |
|
Python |
265 |
Efficient Inference for Big Models |
Jun 24, 2022 |
|
Python |
33 |
llama onnx models and onnxruntime demo |
Apr 09, 2023 |
|
C# |
17 |
.NET wrapper for LLaMA.cpp for LLaMA language model inference on CPU. 🦙 |
Apr 09, 2023 |
|
Python |
2 |
Inference Codes for LLaMA with Intel Extension for Pytorch (Intel Arc GPU) |
Mar 29, 2023 |
|
Python |
10 |
Code for the paper "Stochastic Variational Inference for Dynamic Correlated Topic Models" |
Dec 23, 2021 |
|
C |
16 |
Includes Code for Inference and Evaluation of Topic Models for Selectional Preferences |
Dec 01, 2021 |
|
Python |
4 |
Optimize PyTorch Models for faster inference |
Jun 22, 2022 |
|
C++ |
781 |
Fast inference engine for Transformer models |
Apr 25, 2023 |
|
C++ |
30 |
Scalable inference for Correlated Topic Models |
Oct 12, 2020 |
|
Rust |
2578 |
Run LLaMA inference on CPU, with Rust 🦀🚀🦙 |
Apr 09, 2023 |
|
Python |
432 |
Run LLaMA (and Stanford-Alpaca) inference on Apple Silicon GPUs. |
Apr 09, 2023 |
|
C |
21 |
Fast inference of Instruct tuned LLaMa on your personal devices. |
Apr 02, 2023 |
|
Rust |
3 |
Run LLaMA inference on CPU, with Rust 🦀🚀🦙 |
Apr 08, 2023 |
|
Rust |
10 |
Run LLaMA inference on CPU, with Rust 🦀🚀🦙 |
Apr 12, 2023 |
|
Python |
1338 |
LLaMA: Open and Efficient Foundation Language Models |
Apr 09, 2023 |
|
Python |
12 |
LLaMA: Open and Efficient Foundation Language Models |
May 29, 2023 |