add batch_size
This commit is contained in:
parent
cc60fce612
commit
b36c680b3f
@ -8,4 +8,4 @@ GEMMA_LORA_PATH = os.getenv("GEMMA_LORA_PATH")
|
|||||||
BGE_MODEL_PATH = os.getenv("BGE_MODEL_PATH")
|
BGE_MODEL_PATH = os.getenv("BGE_MODEL_PATH")
|
||||||
BGE_LORA_PATH = os.getenv("BGE_LORA_PATH")
|
BGE_LORA_PATH = os.getenv("BGE_LORA_PATH")
|
||||||
|
|
||||||
BATCH_SIZE = 250
|
BATCH_SIZE = 100
|
||||||
@ -55,6 +55,7 @@ def embed_gemma(request: EmbedRequest):
|
|||||||
raise HTTPException(status_code=400, detail="Invalid model")
|
raise HTTPException(status_code=400, detail="Invalid model")
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
torch.cuda.empty_cache()
|
||||||
raise HTTPException(status_code=500, detail=str(e))
|
raise HTTPException(status_code=500, detail=str(e))
|
||||||
|
|
||||||
torch.cuda.empty_cache()
|
torch.cuda.empty_cache()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user