I like your thinking!
For me, learning from sentences is supremely helpful for learning read, listen, write, and speak.
That said, if the sentences aren’t translated/curated in advance, it is extremely difficult and time-consuming to
1)gather the sentences in Chinese (Mandarin) and
2)ACCURATELY translate them into English.
The most robust source of new sentences for me are subtitles, especially on popular, well-known streaming services that offer subtitle support for all the major languages. This is something that was certainly not available on a wide scale say 20, even 10 years ago, but it certainly is now.
The beauty of subtitles is that the exact same sentences are available in a completely matched set in Chinese (Mandarin) and English or your target native language.
These subtitles are curated for broadcast to subscribers and therefore accurate and reliable enough for learning.
I know how to save the screenshots of the subtitles and copy and paste the text from the images but this is extremely time consuming and inefficient. I’ve stubbornly tried and that’s definitely not the way to go.
As I’ve mentioned before, I’m a beginning computer science/IT student at university, but with my current basic knowledge/skill set I do not know how to go about finding where in streaming service platforms the subtitles are stored or generated from.
The ability to intersect and capture that data in text format would be a monumental, force-multiplying game changer.
It could easily be imported into a text file, spreadsheet, and manipulated rapidly for importing into an SRS platform.
All these steps can certainly be programmed.
Imagine being able to efficiently mine sentences en masse from your absolute favorite television shows and movies while you watch them and then efficiently learn them with the power of SRS!
Any guidance or direction would be appreciated. I’ll do the work. Just point. Thank you!