| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
|
|
|
| |
https://prtimes.jp/main/html/rd/p/000000032.000007490.html
"radiko to launch new service "Timefree 30" this autumn, allows listening
to programmes from the past 30 days, no 3-hour time limit"
probably can't spoof having the plan, that's fine
it might work if you pass cookies of a timefree30 account though,
so i'm adapting the time stuff to account for that
the plan doesn't exist yet, and i don't know how i would go about
detecting it yet, so i'm just hardcoding to False for now
|
|
|
|
|
|
|
|
|
| |
api ignores end_at_lt, it just sets start_at to the start of the broadcast day
the idea was to only get stuff that's actually finished,
so only stuff where the end is before now
with start_at it'll get ongoing stuff as well,
but that's better than not getting the day at all lol
|
|
|
|
|
|
|
|
|
| |
region wasnt used at all
we can make the token stuff into a dict and include the user as well
...which is what i was already doing in the cache
one day i want to change the cache format to use better var names
but i will deal with that later
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Programmes that air on a network of stations have an originating station-
a "key" station. e.g. for all night nippon the key station is ニッポン放送 (LFR)
radiko's new APIs include a key_station field.
I'd only seen such a field in the mobile-web site until now, so I
haven't been able do anything with it. But now the PC website's gotten
this "persons" API, so I can.
this arg is currently only useful for the new "persons" urls,
but i suspect it might become more useful as time goes on...
example usage:
you give the url for the host of a networked show,
the default will download from every station (probably not what you want)
but if you specify key-station-only, then you get just one download from
the originating station
i may add a "preference" argument at some point, for if you want it from
a specific non-key station, but not right now
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
those urls give a bio and pic of the person, and recently they've
started giving a list of all the programmes that person's appeared in as
well.
the API is a bit annoying, you have to specify a start and end time to
search between, so i had to work out the logic for "everything currently
on timefree"
also, the site doesn't use the "end_at_lt" parameter, only start_at_lt
i tried it in the url on the offchance it would work and save a bit of
logic
it did so i will keep it, but it probably stands out a quite a bit, idk
if the mobile apps use this, or with the end_at_lt parameter
the only playlist details I can provide are the ID. There is an API for
the actual details, but that's an extra request, and it's protobuf.
|
|
|
|
|
| |
so just the one --extractor-args incantation can be used for both live
and timefree (--extractor-args "rajiko:device=pc_html5")
|
|
|
|
| |
generally we should work in datetimes, the strings should be the "special" ones
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
| |
all tests passed as of Sat, 03 Aug 2024 06:44:00 +0100
updated to present day where necessary
rearranged the dicts to make them easier to read + modify
'single quotes' -> "double quotes" (including some other bits i saw along the way)
got rid of most of the timefree tests, since the time handling is tested externally now
the station websites stopped using the stationbutton (or started rendering it with javascript) so now cocolo is the only webpage test
|
|
|
|
| |
the link wasnt dead after all
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
wanted to do this before, but per comment:
the official docs led me to believe that importlib_resources was the only
way to do it, and that the thing i was trying to do wasn't supported in
some python versions
so i put it off to one side for later
later came (now), and i tried to implement it with importlib.resources
didn't work, i looked it up and found this stackoverflow post
https://stackoverflow.com/a/58941536
and the second answer, using pkgutil (not importlib.resources), just worked
and the syntax/api seems much nicer as well
have tested with pip yt-dlp + raw directory plugin,
and binary yt-dlp + .whl plugin
also on windows, with binary+whl
both worked, amazing success! so pkgutil seems like the way to go
i also updated the build thing to include .bin files in the extractor dir
|
|
|
|
| |
parity with yt-dlp proper
|
| |
|
| |
|
|
|
|
| |
now links to the actual issue page and prints info useful for debugging
|
|
|
|
|
| |
it does not need to be cryptographically strong
only using token_hex for convenience
|
|
|
|
|
|
|
|
|
|
| |
also fixed a rather embarrassing bug
the generated user agent was, for example:
Dalvik/2.1.0 (Linux; U; Android {'sdk': '27', 'builds': ['5794017', '6107733', '6037697']};SOV36/6107733)
because i put the wrong variable name in the fstring
thats alright its only been sending botched identification for
its entire existence(!)
|
|
|
|
|
| |
makes it a bit easier to use in the cli since you don't need to 'quote'
for the #!
|
|
|
|
| |
there is a live format that also doesn't work
|
|
|
|
| |
closes #18
|
|
|
|
|
|
|
|
|
|
|
| |
so now it's just datetime.datetime with custom input and a few new funcs
instead of actually having a datetime.datetime internally and jankily
kind-of-exposing it
doesn't really affect things currently, but it makes something i want to
do possible (and it's generally just the better way of doing it)
Co-Authored-By: 9382 <omnomaidan@gmail.com>
|
| |
|
|
|
|
|
| |
housou daigaku is going away soon, so switching to radio nikkei
https://www.ouj.ac.jp/news/2023/information/5-32-2.html
|
|
|
|
| |
mostly whitespace changes
|
| |
|
|
|
|
|
| |
extractor shouldnt be executable
time thing should so i can test it easier
|
|
|
|
| |
closes #17
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
Station metadata changes extremely rarely, if ever, so it seems awfully
silly to keep downloading the exact same metadata every time we extract.
Wastes requests, wastes time (300ms round trip for me)
This commit makes it so station metadata is cached for a day, and cached
data is used in place of downloaded metadata when available.
Closes #16
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
The old behaviour assumed programmes were deleted precisely a week after
their end time. This isn't actually the case though, as long as it's
within a week of the _broadcast day_, the site will let you.
This lead to the extractor bailing out (Programme is no longer available)
when in fact the programme was still playable on the site.
The fix uses the same logic as RadikoTime.broadcast_day to find the
broadcast day (TODO: one func for both?), then sets the time to 05:00:00
the next day - i.e. the start of a new broadcast day.
|
| |
|
| |
|
|
|
|
|
| |
now only one thing gets passed around and it has most everything we need
closes #11
|
| |
|
|
|
|
| |
ie dont do the dumb noformats vars, just raise the message there and then
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
afaict nhk doesn't have timefree, and it says as much on the programme pages
The download doesn't fail though, you just get a loop of elevator music
and an apology that it's not available (lasting the duration of the programme)
from tsDetail.js
tsInNg = playable in the area
tsOutNg = playable outside the area
(maybe - just a guess really, but seems to line up)
outside station area + playable in area + not playable outside = not available in your area
outside station area + not playable in area + not playable outside area = not available on timefree
inside station area + not playable in area = not available on timefree
we'll always be in the station area, so we only have to check tsInNg
included a message about NHK Radiru because so far i've only seen NHK programmes
be unavailable
one day i should make the programme_meta func return a dict or something
instead of just adding stuff on the end ad infinitum
closes #15
|
|
|
|
|
|
|
|
|
| |
fixes #14
new headers thing made b'bytes' into "b'bytes'", instead of just 'bytes'
so it sent eg `X-Radiko-Partialkey: b'9lOaAg+2TZRSF+IsiDnUJg=='`
likely a fix for this upstream soon, but will do the sensible thing and make it ascii myself
https://discord.com/channels/807245652072857610/1112613156934668338/1134879336512634880 (yt-dlp discord)
|
| |
|
|
|
|
|
| |
ERROR: Response.info() is deprecated, use Response.headers
see #14
|