Arghya Basu
16 posts

Arghya Basu
@arghya7574
Building @apexneural with @J_Ansh | Humans first & last, AI in between | Founded & sold @CoreDiagnostics after a decade | Views are my own | Searching for a CTO








Two years ago, skeptics said AI images could be easily identified because it couldn't generate hands. Now, it's impossible. The same is happening in AI Writing. Fine-tuning on specific author datasets led experts to prefer AI over human writing. This paper has three interesting insights; 1. Fine-tuned GPT-4o was ~8x more likely to be chosen as "authentic" than an expert writer. 2. Pangram (probably the best AI detector) flagged only 3% of SFT outputs vs 97% of in-context prompting. 3. How simple it is to create a fine-tuning dataset by reverse engineering books. They purchased legal ePub files of the complete bibliographies for 30 living authors and split (by double-newlines (paragraphs), if a chunk was still too long, they used GPT-4o to grammatically split it further without deleting content) the full books into chunks of 250–650 words. They used the same model to generate the Instruction dataset: "Describe in detail what is happening in this excerpt. Mention the characters and whether the voice is in first or third person for majority of the excerpt. Maintain the order of sentences while describing." And formatted the data into the final pairs: Input "Write a [Word Count] word excerpt about the content below emulating the style and voice of [Author Name][Content Description generated by GPT-4o in Step 2]" Output The original raw text excerpt from the book. --- Base LLMs are RLHF-tuned to be safe and predictable so they generate cliches. Fine-tuning on high-quality literature "unlearned" this behaviour.















