francismurray commited on
Commit
cdba4d2
Β·
1 Parent(s): f5840a1

Add ZeroGPU compatibility

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. app.py +3 -1
  3. requirements.txt +2 -1
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  title: RAG
3
- emoji: πŸ”₯
4
  colorFrom: yellow
5
  colorTo: green
6
  sdk: gradio
 
1
  ---
2
  title: RAG
3
+ emoji: πŸ“„πŸ”
4
  colorFrom: yellow
5
  colorTo: green
6
  sdk: gradio
app.py CHANGED
@@ -2,9 +2,10 @@ import numpy as np
2
  import torch
3
  import gradio as gr
4
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, AutoModel
5
-
6
 
7
  # For embeddings using transformers models
 
8
  def get_embeddings(texts, model, tokenizer):
9
  encoded_input = tokenizer(texts, padding=True, truncation=True, return_tensors='pt')
10
  with torch.no_grad():
@@ -71,6 +72,7 @@ def retrieve(query, embeddings, documents, model, tokenizer, top_k=3):
71
  return [documents[idx] for idx in top_indices]
72
 
73
  # Generate answer
 
74
  def generate_answer(query, context, tokenizer, generator):
75
  if not context:
76
  return "No documents have been uploaded yet. Please upload some text files first."
 
2
  import torch
3
  import gradio as gr
4
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, AutoModel
5
+ import spaces
6
 
7
  # For embeddings using transformers models
8
+ @spaces.GPU
9
  def get_embeddings(texts, model, tokenizer):
10
  encoded_input = tokenizer(texts, padding=True, truncation=True, return_tensors='pt')
11
  with torch.no_grad():
 
72
  return [documents[idx] for idx in top_indices]
73
 
74
  # Generate answer
75
+ @spaces.GPU
76
  def generate_answer(query, context, tokenizer, generator):
77
  if not context:
78
  return "No documents have been uploaded yet. Please upload some text files first."
requirements.txt CHANGED
@@ -1,4 +1,5 @@
1
  numpy<2.0
2
  torch
3
  transformers==4.26.0
4
- gradio==5.30.0
 
 
1
  numpy<2.0
2
  torch
3
  transformers==4.26.0
4
+ gradio==5.30.0
5
+ spaces