TLDR:
I have built a custom GPT: What Would Jobs Do?
https://chatgpt.com/g/g-68f7a59494c481918995e7eecb050617-what-would-jobs-do-wwjd
Long Story:
When I was 11 or 12, a neighborhood senior kid told me about Apple, Mac and Steve Jobs. I genuinely thought he was bullshitting me thinking that i don’t know about Bill Gates.
Fast forward, when I wanted to buy a gift for someone very close during my first onsite trip, a friend was like, if you really want to gift properly, get them a iPod mini. It was the entry product and still it was too expensive when compared to other mp3 players.
I was curious and went down the rabbit hole and that is how my fascination and adoration with Steve Jobs got started. It was too strong, I ended up writing a book about him in Tamil. It is been more than 10 years and now and then I still get fan mail for it.
The unintended consequence of deep diving about someone like Steve Job is, me losing my mind over when people (read tech bros, founders) channel their inner “A” and passing it as modeling the Jobsian behavior.
Probably, one of the most misunderstood personality. Thanks to the way, he was portrayed by most of main stream media (For what it’s worth, folks close to Steve himself felt the authorized biography of Walter Isaacson itself didn’t do justice and they got Brent Schlender and another author for redux of Steve Jobs bio)
So for more than few years, I have been meaning to write two books in English.
One, compile all his quotes and put it as a book.
Two, based on quotes, based on pivotal moments, write about how one could learn from Jobs to make a dent in the universe rather than being a painful “A”
I started compiling all his videos, interviews, books that I missed earlier and the books published after my book. (I had to rely a lot on iCon and lot of early articles across publications like Forbes, BusinessWeek, Time, Wired and Fortune to write my book; The Walter Isaacson wasn’t published when I was done with mine)
However, I never got around to writing them. I couldn’t get around to getting the transcripts from the YouTube video complilations.
Later, I started playing around with NotebookLM and chatGPT to write scripts to compile the transcripts from videos. Meanwhile, I also wanted to be AI literate and was reading around and playing around with tools.
Then it occured to me, wait a minute, what if I use AI to model Steve Jobs behavior. I have these many books, these many articles, videos and I could give it Open AI’s GPT builder. Simple. (I couldn’t be more correct as well as more wrong; Correct cause, it is still relatively easy to use the GPT builder but still it is not that simple to build the Knowledge files to serve as inputs to the GPT builder)
One, using the books as a whole would violate, Open AI’s usage guidelines which prioritizes fair use boundaries.
More than that, the Input Token limits and practicality.
The GPT file upload system doesn’t “absorb” the entire content at once like training a model. Instead, it uses retrieval techniques that pull in relevant file snippets (based on the prompt) when generating a response.
- Even If you upload an entire 700-page book, or books it won’t be searched efficiently.
- You may get partial context, or the model may miss relevant insights buried deep in long documents.
- There’s also a token cap (~100k per request) which limits how much of the file it can process per response.
Feeding a GPT unstructured, untagged, long-form books:
- Makes the assistant guess context too often.
- Reduces precision in responses.
So, I realized this was more challenging (in a way) than writing a book cause, you still have to create topic notes from books, themes, organize it and not just that, these files should explicitly call out as the GPT wouldn’t automatically “know” what’s from Jobs and what’s from me unless I make that distinction explicit in the formatting. It needs, content in a specific format, with attributes and tags, calling out, Quote, Source, Comment, Interpretation and the files cannot be ambiguous or cannot have unclear content. GPT may not know where the quote ends and where my comment begins. It might conflate my interpretation with Jobs’ own words.
It is cause of the way, how the GPT works.
When you upload a file to a Custom GPT, it doesn’t “read” the whole thing every time. Instead, it:
- Searches for the most relevant chunks of content related to the prompt.
- Pulls those chunks into the response as background.
- Uses context to decide what’s a quote, what’s commentary, and how to blend it into the answer.
The clearer your formatting, the more accurate this retrieval. So how to build the files.
Q1:.txt or .docx — Which Is Better?
|
Format |
Pros |
Cons |
|
.txt |
” Lightweight, loads fast✔ No hidden formatting✔ GPT parses text cleanly |
No headings, styles, or tables Can get messy with long documents |
|
.docx |
Supports headers, bolding, bullet points, tables✔ Good for structured documents with commentary |
Slightly heavier Risk of hidden metadata (tracked changes, formatting issues) |
Recommendation:
- If you’re uploading clean, well-structured material with multiple sections, .docx is better.
- If you’re uploading shorter, raw text content like quotes or notes, .txt is fine.
- In both formats, use clear headers or tags ([QUOTE], [COMMENT], etc.) for GPT to parse meaningfully.
You can use .docx files with tables, and GPT can infer structure better from them than from unformatted .txt.
One Big File vs. Multiple Small Files?
|
Approach |
Pros |
Cons |
|
One Large File |
Easy to manage✔ Avoids GPT file upload limit (“10 files max)Good for a single tightly themed topic |
Retrieval becomes fuzzy if file mixes many unrelated topics⌠Hard to update incrementally GPT may struggle to find the most relevant snippet |
|
Multiple Small Files |
Better for topic-based retrieval” Easier to test and update individually” More accurate context matching |
More upload effort (GPT Builder has file count limits) Slightly harder to maintain if not organized well |
Why this works:
- GPT fetches only what’s needed based on your prompt (e.g., hiring questions won’t trigger irrelevant product quotes)
- Easier for you to add/update one theme without breaking the others
- Each file must be ≤ 512 MB .
- Text/doc files are capped at 2 million tokens per file
- GPT indexes each file separately
- When a user asks a question (e.g., “What would Jobs say about launching MVPs?”), it:
- Searches for relevant files (e.g., Design + Product + Risk files)
- Extracts relevant quotes/snippets
- Uses those to inform the response
➡️ Better-separated files = better semantic matches, fewer irrelevant quotes, and faster processing.