| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
| |
still hardcoded to not work though
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
https://prtimes.jp/main/html/rd/p/000000032.000007490.html
"radiko to launch new service "Timefree 30" this autumn, allows listening
to programmes from the past 30 days, no 3-hour time limit"
probably can't spoof having the plan, that's fine
it might work if you pass cookies of a timefree30 account though,
so i'm adapting the time stuff to account for that
the plan doesn't exist yet, and i don't know how i would go about
detecting it yet, so i'm just hardcoding to False for now
|
|
|
|
|
|
|
|
|
| |
api ignores end_at_lt, it just sets start_at to the start of the broadcast day
the idea was to only get stuff that's actually finished,
so only stuff where the end is before now
with start_at it'll get ongoing stuff as well,
but that's better than not getting the day at all lol
|
|
|
|
|
|
|
|
|
| |
region wasnt used at all
we can make the token stuff into a dict and include the user as well
...which is what i was already doing in the cache
one day i want to change the cache format to use better var names
but i will deal with that later
|
| |
|
| |
|
|
|
|
|
| |
changed func in eb94441420845970d48fe7b4e30cc5eed78ac95a , forgot to
change the test also
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Programmes that air on a network of stations have an originating station-
a "key" station. e.g. for all night nippon the key station is ニッポン放送 (LFR)
radiko's new APIs include a key_station field.
I'd only seen such a field in the mobile-web site until now, so I
haven't been able do anything with it. But now the PC website's gotten
this "persons" API, so I can.
this arg is currently only useful for the new "persons" urls,
but i suspect it might become more useful as time goes on...
example usage:
you give the url for the host of a networked show,
the default will download from every station (probably not what you want)
but if you specify key-station-only, then you get just one download from
the originating station
i may add a "preference" argument at some point, for if you want it from
a specific non-key station, but not right now
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
those urls give a bio and pic of the person, and recently they've
started giving a list of all the programmes that person's appeared in as
well.
the API is a bit annoying, you have to specify a start and end time to
search between, so i had to work out the logic for "everything currently
on timefree"
also, the site doesn't use the "end_at_lt" parameter, only start_at_lt
i tried it in the url on the offchance it would work and save a bit of
logic
it did so i will keep it, but it probably stands out a quite a bit, idk
if the mobile apps use this, or with the end_at_lt parameter
the only playlist details I can provide are the ID. There is an API for
the actual details, but that's an extra request, and it's protobuf.
|
|
|
|
|
| |
so just the one --extractor-args incantation can be used for both live
and timefree (--extractor-args "rajiko:device=pc_html5")
|
|
|
|
| |
generally we should work in datetimes, the strings should be the "special" ones
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
| |
all tests passed as of Sat, 03 Aug 2024 06:44:00 +0100
updated to present day where necessary
rearranged the dicts to make them easier to read + modify
'single quotes' -> "double quotes" (including some other bits i saw along the way)
got rid of most of the timefree tests, since the time handling is tested externally now
the station websites stopped using the stationbutton (or started rendering it with javascript) so now cocolo is the only webpage test
|
|
|
|
| |
the link wasnt dead after all
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
wanted to do this before, but per comment:
the official docs led me to believe that importlib_resources was the only
way to do it, and that the thing i was trying to do wasn't supported in
some python versions
so i put it off to one side for later
later came (now), and i tried to implement it with importlib.resources
didn't work, i looked it up and found this stackoverflow post
https://stackoverflow.com/a/58941536
and the second answer, using pkgutil (not importlib.resources), just worked
and the syntax/api seems much nicer as well
have tested with pip yt-dlp + raw directory plugin,
and binary yt-dlp + .whl plugin
also on windows, with binary+whl
both worked, amazing success! so pkgutil seems like the way to go
i also updated the build thing to include .bin files in the extractor dir
|
|
|
|
| |
parity with yt-dlp proper
|
| |
|
| |
|
|
|
|
| |
now links to the actual issue page and prints info useful for debugging
|
|
|
|
|
| |
it does not need to be cryptographically strong
only using token_hex for convenience
|
|
|
|
|
|
|
|
|
|
| |
also fixed a rather embarrassing bug
the generated user agent was, for example:
Dalvik/2.1.0 (Linux; U; Android {'sdk': '27', 'builds': ['5794017', '6107733', '6037697']};SOV36/6107733)
because i put the wrong variable name in the fstring
thats alright its only been sending botched identification for
its entire existence(!)
|
|
|
|
|
| |
makes it a bit easier to use in the cli since you don't need to 'quote'
for the #!
|
|
|
|
| |
there is a live format that also doesn't work
|
|
|
|
| |
closes #18
|
|
|
|
|
|
|
|
|
|
|
| |
so now it's just datetime.datetime with custom input and a few new funcs
instead of actually having a datetime.datetime internally and jankily
kind-of-exposing it
doesn't really affect things currently, but it makes something i want to
do possible (and it's generally just the better way of doing it)
Co-Authored-By: 9382 <omnomaidan@gmail.com>
|
| |
|
|
|
|
|
| |
housou daigaku is going away soon, so switching to radio nikkei
https://www.ouj.ac.jp/news/2023/information/5-32-2.html
|
|
|
|
| |
mostly whitespace changes
|
| |
|
|
|
|
|
|
| |
deduped code for working it out
also fixed bug in broadcast_day_end where it shifted forward by a day,
but didn't also set the time to 5am
|
|
|
|
|
|
|
|
|
| |
the site will accept this so we have to as well
i should probably try and handle this in SiteTime
but cant be fucked right now
iirc the site just blanks out if the day is out of range
so the datetime exception matches the site's behaviour 👍
|
|
|
|
|
| |
extractor shouldnt be executable
time thing should so i can test it easier
|
|
|
|
| |
closes #17
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
Station metadata changes extremely rarely, if ever, so it seems awfully
silly to keep downloading the exact same metadata every time we extract.
Wastes requests, wastes time (300ms round trip for me)
This commit makes it so station metadata is cached for a day, and cached
data is used in place of downloaded metadata when available.
Closes #16
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
The old behaviour assumed programmes were deleted precisely a week after
their end time. This isn't actually the case though, as long as it's
within a week of the _broadcast day_, the site will let you.
This lead to the extractor bailing out (Programme is no longer available)
when in fact the programme was still playable on the site.
The fix uses the same logic as RadikoTime.broadcast_day to find the
broadcast day (TODO: one func for both?), then sets the time to 05:00:00
the next day - i.e. the start of a new broadcast day.
|
| |
|
| |
|
|
|
|
|
| |
now only one thing gets passed around and it has most everything we need
closes #11
|
| |
|
|
|
|
| |
ie dont do the dumb noformats vars, just raise the message there and then
|