Text Generation

View as Markdown

Harness the power of large language models to generate any type of text content - from creative writing and code to structured data formats and technical documentation.

Basic Usage

Generate text responses using our chat completion endpoint:

1from openai import OpenAI
2
3client = OpenAI(
4 base_url="https://api.neura-ai.app/v1"
5)
6
7response = client.chat.completions.create(
8 model="mistral-medium-latest",
9 messages=[
10 {"role": "user", "content": "Explain quantum computing in simple terms"}
11 ],
12 temperature=0.7
13)
14
15print(response.choices[0].message.content)

Controlling Output Style

Temperature

Adjust the temperature parameter to control creativity vs. consistency:

1# More focused and deterministic
2response = client.chat.completions.create(
3 model="gpt-5",
4 messages=[{"role": "user", "content": "What is 2+2?"}],
5 temperature=0.1
6)
7
8# More creative and varied
9response = client.chat.completions.create(
10 model="gpt-5",
11 messages=[{"role": "user", "content": "Write a creative story opening"}],
12 temperature=1.5
13)

System Messages

Guide the model’s behavior with system instructions:

1response = client.chat.completions.create(
2 model="gpt-5",
3 messages=[
4 {
5 "role": "system",
6 "content": "You are a helpful coding assistant specializing in Python"
7 },
8 {
9 "role": "user",
10 "content": "How do I read a CSV file?"
11 }
12 ]
13)

Use Cases

Code Generation

1response = client.chat.completions.create(
2 model="gpt-5",
3 messages=[{
4 "role": "user",
5 "content": "Write a Python function to calculate fibonacci numbers with memoization"
6 }],
7 temperature=0.2
8)

Content Summarization

1long_article = "..." # Your long text here
2
3response = client.chat.completions.create(
4 model="mistral-medium-latest",
5 messages=[{
6 "role": "user",
7 "content": f"Provide a concise summary of this article:\n\n{long_article}"
8 }],
9 temperature=0.3
10)

Structured Output

Generate JSON or other structured formats:

1response = client.chat.completions.create(
2 model="gpt-5",
3 messages=[{
4 "role": "user",
5 "content": "Generate a JSON object with 5 fake user profiles including name, email, and age"
6 }],
7 temperature=0.7
8)

Streaming Responses

For real-time output, enable streaming:

1stream = client.chat.completions.create(
2 model="gpt-5",
3 messages=[{"role": "user", "content": "Write a poem about the ocean"}],
4 stream=True
5)
6
7for chunk in stream:
8 if chunk.choices[0].delta.content is not None:
9 print(chunk.choices[0].delta.content, end="")

Best Practices

  • Use lower temperatures (0.1-0.3) for factual or deterministic tasks
  • Use higher temperatures (0.7-1.2) for creative or varied outputs
  • Include clear, specific instructions in your prompts
  • Use system messages to set consistent behavior
  • Consider token limits when working with large inputs or outputs