Comment by lovestory
My personal workflow with memorizing with Anki using LLMs is as follows: Read textbook material first. It's important that you understand what you are learning.You cannot just breakdown information into atoms and expect to understand how it works together (for example: you can learn that there are monounsaturated fats, polyunsaturated fats, saturated fats and trans fats. But without reading beforehand about them from a textbook or other source, you will not understand how they differ (in chemical structure, biological function etc). After I understand the material, I feed LLM the documents (textbooks etc) and give it the following prompt: >I want to generate flashcards from the provided textbook document using the attached PDF. Each flashcard should contain a question and a corresponding answer formatted as pairs in plaintext code block. The structure should be: "Question","Answer" >Extract key concepts, definitions, and explanations from the textbook. If in the text imperial unit system is used, convert it to metric. For mathematical symbols and equations, format them using inline MathJax syntax because I will be importing the copied text to Anki. Ensure questions are clear and concise while answers provide a direct yet comprehensive response.
After the text is generated, I check out the accuracy (in 95% of cases the cards are accurate) and I import them into my decks. The rest is good old school Anki memorizing.
I remember trying that a few months back but I was not convinced by the quality of the cards. It's also hard to convey what I think a good card is, I guess there's something about not getting lost in the details while missing the big picture.
You only commented on accuracy, but what's your experience on relevance and how useful LLM-generated flashcards are?
To be clear I've already found myself deleting some flashcards I made myself while reviewing them when I realized they were bad, so I guess one can do that for LLM-generated questions as well, as long as the irrelevance rate is somewhat similar.