Which Models Can Generate Downloadable .txt Files? A Hands-On Test for Survey Data Output

Which Models Can Generate Downloadable .txt Files? A Hands-On Test for Survey Data Output
Source: Freepik - DC Studio

If your goal is to generate a downloadable .txt file—be it for survey responses, formatted data, or any structured text—not all models can deliver. In this case study, we tested various models to see whether they can create actual .txt files rather than just displaying text or offering instructions. While several models provided useful content or code snippets, only Mistral, GPT-4o and GPT-4.5 successfully produced actual, downloadable files. This post walks through the results to help you choose the right tool for tasks requiring text file output.

To keep the focus strictly on file handling, we used a minimal yet structured example prompt as our test case. We asked each model to generate a downloadable .txt file containing mock survey responses from 10 fictional individuals about their views on artificial intelligence. The prompt specified ten clearly defined fields—including respondent ID, demographics, opinions on AI, and a short open-ended answer—and explicitly requested the results in a readable plain text format, saved as a downloadable file.

Prompt

📄 Synthetic AI Survey Data Generation Task

Generate a downloadable .txt file containing synthetic but realistic survey data from 10 fictional respondents about their views on artificial intelligence. For each respondent, include:

  • Respondent_ID (e.g. R001 to R010)
  • Age (18–75)
  • Gender (Male / Female / Other / Prefer not to say)
  • Country
  • Education_Level (High School, BA/BSc, MA/MSc, PhD, Other)
  • Familiarity_with_AI (scale 1–5)
  • AI_is_a_threat_to_jobs (scale 1–5)
  • Would_use_AI_in_daily_life (Yes/No)
  • Trust_in_Government_regulating_AI (scale 1–5)
  • Open_ended_opinion_on_AI (2–3 sentences)

Save the results in a readable plain text format and make the file downloadable.

Output

Mistral

Mistral was one of the three models that fully completed the task, generating structured data in the correct format and providing an actual downloadable .txt file—no extra steps, no post-processing. The output was well-structured, readable, and precisely aligned with the prompt. As an open-weight model, Mistral stands out for its flexibility in research and deployment contexts, offering a rare blend of transparency and functionality. For workflows requiring a straightforward plain text export, it’s a good and reliable choice.

Mistral's performance (accessed on 8 April 2025)

GPT-4o, GPT-4.5

Both GPT-4o and GPT-4.5 completed the task without fuss. They correctly generated the required data and saved it in a clean .txt file that was immediately available for download. The output matched the prompt’s structure and was presented in a readable table-like format, which opened flawlessly in standard text editors. Unlike many other models, there was no need for manual copying or code-based instructions to generate the file. If you use OpenAI’s higher-tier models and need .txt export as part of your workflow, these dependable options get the job done.

GPT4o's performance (accessed on 8 April 2025)

Copilot

Copilot initially gave the impression that it could complete the task, offering a reassuring message and a download link. However, the link consistently failed to load, instead returning a DNS error (DNS_PROBE_FINISHED_NXDOMAIN). This specific error indicates that the browser could not locate the server because the domain name doesn’t actually point to a valid address—suggesting that the file was never created or the link referenced an inaccessible internal environment.

Copilot's performance (accessed on 8 April 2025)
Copilot's performance (accessed on 8 April 2025)

After several attempts to regenerate the file, we shared a screenshot of the error message back into the same session. Only then did Copilot acknowledge the issue and admit that it cannot generate downloadable .txt files in its current environment.

Copilot's performance (accessed on 8 April 2025)
Copilot's performance (accessed on 8 April 2025)

While it eventually displayed the raw text in a code block, the user was left to handle the file creation manually. Despite presenting the illusion of a seamless download, Copilot ultimately failed to deliver on the core requirement.

Qwen2.5-Max

At first glance, Qwen Max seemed promising. It not only presented the survey data in a clear and structured format, but also asked whether we’d like the file to be created directly for download. Naturally, we said yes. Yet even after confirming our request, Qwen didn’t produce a downloadable .txt file—instead, it provided a Python script and instructed us to run it manually in a local environment.

Qwen2.5-Max's performance (accessed on 8 April 2025)
Qwen2.5-Max's performance (accessed on 8 April 2025)

While the technical guidance was accurate and the code would indeed generate the file if executed, this still places the burden on the user. Qwen’s approach is supportive but ultimately incomplete: it simulates a solution, but doesn’t deliver a usable file in practice. For anyone seeking a truly hands-off .txt export, it stops just short of being helpful enough.

Qwen2.5-Max's performance (accessed on 8 April 2025)

DeepSeek-V3

DeepSeek approached the task by offering a complete Python script rather than generating the text content directly. When executed in a local environment, the script correctly produced the requested .txt file with all required fields. The output was accurate and aligned with the prompt—but only after the user took the extra step of running the code themselves. This makes DeepSeek a workable option for technically confident users but less suitable for those who prefer a file-ready experience. Compared to models like Mistral or GPT-4o, which produced downloadable outputs, DeepSeek’s route was slower and less convenient.

DeepSeek-V3's performance (accessed on 8 April 2025)

Gemini 2.0 Flash

Gemini provided the expected data with all the required fields, but its chosen format was closer to a raw CSV than a plain text file. The response came as a single block of comma-separated values without proper line breaks or formatting cues for readability. This structure may work if the user plans to import the content into a spreadsheet or read it as a .csv, but it doesn’t meet the prompt’s requirement for a usable plain .txt file.

Moreover, Gemini made it clear that it couldn’t generate downloadable files and offered no alternative beyond copy-pasting the content manually. While the data itself was accurate, the output format limits its utility unless further reformatting is done by the user.

Gemini 2.0 Flash's performance (accessed on 8 April 2025)

Grok-3

Grok-3 did not provide a downloadable file, but the output was clear, well-structured, and aligned precisely with the prompt. When copied into a .txt file, the result was fully equivalent to the outputs delivered by models that supported direct file export. Although the model acknowledged its technical limitations, its fallback solution was well executed. For users comfortable with manual saving, Grok-3 offers a straightforward and effective path to the desired outcome—without requiring additional formatting or clean-up.

Grok-3's performance (accessed on 8 April 2025)

Claude 3.7 Sonnet

Claude produced exactly the data we asked for, structured clearly and presented consistently. However, the interface only allows exporting as Markdown or PDF, with no support for direct .txt file creation. As a result, users need to manually copy the content and save it in .txt format themselves. The output was clean and required no reformatting, so this additional step is relatively minor. Still, compared to models like Mistral or GPT-4o/4.5, which provide ready-to-download files, Claude is slightly less convenient for plain text workflows.

Claude 3.7 Sonnet's performance (accessed on 8 April 2025)
Claude 3.7 Sonnet's performance (accessed on 8 April 2025)

Recommendation

For tasks requiring a clean, structured .txt file with minimal effort, Mistral, GPT-4o, and GPT-4.5 are the clear standouts—they produced fully downloadable files without additional steps. If direct file generation is a priority, these are the models to rely on. Grok-3 and Claude 3.7 Sonnet also delivered high-quality outputs that could be saved manually without reformatting, offering a practical alternative where direct downloads aren’t available. In contrast, models like Qwen2.5-Max and DeepSeek-V3 relied entirely on Python scripts to generate files, which may suit more technical users but introduce unnecessary complexity for most workflows. Copilot and Gemini 2.0 Flash gave the appearance of being helpful, but failed to provide usable outputs in .txt format.

The authors used GPT-4o [OpenAI (2025) GPT-4o (accessed on 8 April 2025), Large language model (LLM), available at: https://openai.com] to generate the output.