Skip to content

Commit 30e9e94

Browse files
committed
docs: add readme
1 parent 9a28f81 commit 30e9e94

File tree

1 file changed

+42
-0
lines changed

1 file changed

+42
-0
lines changed

README.md

+42
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,42 @@
1+
# FastAPI Llama2 Huggingface Hub API
2+
3+
This repository contains a Jupyter notebook (`FastAPI-Llama-HuggingfaceHub-Collab.ipyn`) that demonstrates how to set up and run a FastAPI server with Llama 2 model integration using Google Colab's free T4 GPU.
4+
5+
## Features
6+
7+
- Sets up a FastAPI server with Llama 2 model integration
8+
- Uses Google Colab's free GPU for model inference
9+
- Creates a public URL for the API using ngrok
10+
- Provides an example of how to make API requests to the server
11+
12+
## Contents
13+
14+
The notebook includes the following main sections:
15+
16+
1. Installation of dependencies
17+
2. Setting up ngrok for creating a public URL
18+
3. Creating the FastAPI application
19+
4. Starting the FastAPI server
20+
5. Using ngrok to create a public URL for the server
21+
6. Testing the API with example requests
22+
23+
## Usage
24+
25+
1. Open the `FastAPI-Llama-HuggingfaceHub-Collab.ipynb` notebook in Google Colab
26+
2. Follow the instructions in the notebook to set up and run the server
27+
3. Use the provided ngrok URL to make API requests to the Llama 2 model
28+
29+
## Requirements
30+
31+
- Google Colab account (for free GPU access)
32+
- ngrok account (free tier is sufficient)
33+
34+
## Note
35+
36+
Make sure to shut down the server and ngrok processes when you're done using the notebook to free up resources.
37+
For more detailed instructions and code explanations, please refer to the comments within the notebook.
38+
39+
## License
40+
41+
Copyright © 2024 [tooniez](https://github.com/tooniez).
42+
This project is [MIT](https://github.com/tooniez/fastapi-llama-huggingfacehub-collab/blob/main/LICENSE) licensed.

0 commit comments

Comments
 (0)