temperature parameter is set above 0, the model will likely produce different results each time - this is expected behavior. If you're seeing unexpected differences in the quality completions you receive from Playground vs. the API with
temperature set to 0, there are a few potential causes to consider.
First, check that your prompt is exactly the same. Even slight differences, such as an extra space or newline character, can lead to different outputs.
Next, ensure you're using the same parameters in both cases. For example, the
model parameter set to
gpt-4 will produce different completions even with the same prompt, because
gpt-4 is a newer and more capable instruction-following model.
If you've double-checked all of these things and are still seeing discrepancies, ask for help on the Community Forum, where users may have experienced similar issues or may be able to assist in troubleshooting your specific case.