Spaces:
Lap-AI
/
Runtime error

Reality123b commited on
Commit
731d6b6
·
verified ·
1 Parent(s): c950d42

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -4
app.py CHANGED
@@ -21,12 +21,19 @@ streamer = TextStreamer(generator.tokenizer, skip_prompt=True)
21
 
22
  def generate_response(prompt: str, max_new_tokens: int = 4096):
23
  try:
24
- messages = [{"role": "user", "content": prompt}]
25
- output = generator(messages, max_new_tokens=max_new_tokens, do_sample=False, streamer=streamer)
26
- return output[0]["generated_text"][-1]["content"]
 
 
 
 
 
 
 
27
  except Exception as e:
28
  raise ValueError(f"Error generating response: {e}")
29
-
30
  @app.post("/generate")
31
  async def generate_text(input: ModelInput):
32
  try:
 
21
 
22
  def generate_response(prompt: str, max_new_tokens: int = 4096):
23
  try:
24
+ # Pass the prompt as a simple string, not a chat message list
25
+ output = generator(prompt, max_new_tokens=max_new_tokens, do_sample=False, streamer=streamer)
26
+
27
+ # The output format is different now. We need to extract the response.
28
+ full_text = output[0]["generated_text"]
29
+
30
+ # Remove the original prompt from the start of the response
31
+ if full_text.startswith(prompt):
32
+ return full_text[len(prompt):].strip()
33
+ return full_text
34
  except Exception as e:
35
  raise ValueError(f"Error generating response: {e}")
36
+
37
  @app.post("/generate")
38
  async def generate_text(input: ModelInput):
39
  try: