deepseek-coder-6.7b-instruct-de1

A Delphi/ObjectPascal coding assistant built on deepseek-coder-6.7b-instruct.

13-line FATPIE system prompt, 4096 context. Proven 127/136 on smoketest.

What is this?

This is not a fine-tuned model. It uses the unmodified deepseek-coder-6.7b-instruct weights with a carefully crafted system prompt that activates the model's existing knowledge of Delphi/ObjectPascal conventions.

Six QLoRA fine-tuning attempts on this base model either had zero effect or caused regression. The system prompt alone produces better results than any fine-tuned variant.

How to use

With Ollama

# Download the Modelfile
# Then create the model:
ollama create deepseek-coder:6.7b-instruct-de1 -f Modelfile.de1

# Run it:
ollama run deepseek-coder:6.7b-instruct-de1 "Write a Delphi function that reverses a string"

What the system prompt teaches

  • Memory management: No ARC or GC on Windows. Free every object you create.
  • FATPIE naming: T for types, F for fields, A for parameters, P for pointers, I for interfaces, E for exceptions
  • Code style: Use Result not FunctionName, begin not BEGIN, try/finally for cleanup
  • Platform: Delphi on Windows, MSBuild for builds, TDateTime for dates

Model details

Key finding

A well-crafted system prompt on an unmodified 6.7B model beats QLoRA fine-tuning with 500-1000 curated instruction pairs. The model already knows Delphi โ€” it just needs the right context to access that knowledge.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for wpostma/deepseek-coder-6.7b-instruct-de1

Finetuned
(60)
this model