1
0
Fork 0
mirror of https://github.com/ytdl-org/youtube-dl.git synced 2024-12-28 03:37:59 +00:00
youtube-dl/youtube_dl/extractor/qqmusic.py

229 lines
8.1 KiB
Python
Raw Normal View History

2015-03-17 11:03:29 +00:00
# coding: utf-8
from __future__ import unicode_literals
2015-03-18 05:56:02 +00:00
import random
import time
2015-03-18 06:59:33 +00:00
import re
2015-03-18 05:56:02 +00:00
2015-03-17 11:03:29 +00:00
from .common import InfoExtractor
2015-03-18 17:47:07 +00:00
from ..utils import (
strip_jsonp,
unescapeHTML,
)
2015-03-18 06:59:33 +00:00
from ..compat import compat_urllib_request
2015-03-17 11:03:29 +00:00
class QQMusicIE(InfoExtractor):
IE_NAME = 'qqmusic'
2015-03-17 11:03:29 +00:00
_VALID_URL = r'http://y.qq.com/#type=song&mid=(?P<id>[0-9A-Za-z]+)'
_TESTS = [{
'url': 'http://y.qq.com/#type=song&mid=004295Et37taLD',
'md5': 'bed90b6db2a7a7a7e11bc585f471f63a',
'info_dict': {
'id': '004295Et37taLD',
'ext': 'm4a',
'title': '可惜没如果',
'upload_date': '20141227',
'creator': '林俊杰',
'description': 'md5:d327722d0361576fde558f1ac68a7065',
2015-03-17 11:03:29 +00:00
}
}]
2015-03-18 05:56:02 +00:00
# Reference: m_r_GetRUin() in top_player.js
# http://imgcache.gtimg.cn/music/portal_v3/y/top_player.js
@staticmethod
def m_r_get_ruin():
curMs = int(time.time() * 1000) % 1000
return int(round(random.random() * 2147483647) * curMs % 1E10)
2015-03-17 11:03:29 +00:00
def _real_extract(self, url):
mid = self._match_id(url)
detail_info_page = self._download_webpage(
'http://s.plcloud.music.qq.com/fcgi-bin/fcg_yqq_song_detail_info.fcg?songmid=%s&play=0' % mid,
2015-03-18 06:59:33 +00:00
mid, note='Download song detail info',
errnote='Unable to get song detail info', encoding='gbk')
2015-03-17 11:03:29 +00:00
song_name = self._html_search_regex(
r"songname:\s*'([^']+)'", detail_info_page, 'song name')
publish_time = self._html_search_regex(
r'发行时间:(\d{4}-\d{2}-\d{2})', detail_info_page,
'publish time', default=None)
if publish_time:
publish_time = publish_time.replace('-', '')
2015-03-17 11:03:29 +00:00
singer = self._html_search_regex(
r"singer:\s*'([^']+)", detail_info_page, 'singer', default=None)
lrc_content = self._html_search_regex(
r'<div class="content" id="lrc_content"[^<>]*>([^<>]+)</div>',
detail_info_page, 'LRC lyrics', default=None)
if lrc_content:
lrc_content = lrc_content.replace('\\n', '\n')
2015-03-17 11:03:29 +00:00
2015-03-18 05:56:02 +00:00
guid = self.m_r_get_ruin()
2015-03-17 11:03:29 +00:00
vkey = self._download_json(
'http://base.music.qq.com/fcgi-bin/fcg_musicexpress.fcg?json=3&guid=%s' % guid,
mid, note='Retrieve vkey', errnote='Unable to get vkey',
transform_source=strip_jsonp)['key']
song_url = 'http://cc.stream.qqmusic.qq.com/C200%s.m4a?vkey=%s&guid=%s&fromtag=0' % (mid, vkey, guid)
return {
'id': mid,
'url': song_url,
'title': song_name,
'upload_date': publish_time,
'creator': singer,
'description': lrc_content,
2015-03-17 11:03:29 +00:00
}
2015-03-18 06:59:33 +00:00
2015-03-18 17:47:07 +00:00
class QQPlaylistBaseIE(InfoExtractor):
@staticmethod
def qq_static_url(category, mid):
return 'http://y.qq.com/y/static/%s/%s/%s/%s.html' % (category, mid[-2], mid[-1], mid)
@classmethod
def get_entries_from_page(cls, page):
entries = []
for item in re.findall(r'class="data"[^<>]*>([^<>]+)</', page):
song_mid = unescapeHTML(item).split('|')[-5]
entries.append(cls.url_result(
'http://y.qq.com/#type=song&mid=' + song_mid, 'QQMusic',
song_mid))
2015-03-18 17:47:07 +00:00
return entries
class QQMusicSingerIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:singer'
2015-03-18 06:59:33 +00:00
_VALID_URL = r'http://y.qq.com/#type=singer&mid=(?P<id>[0-9A-Za-z]+)'
_TEST = {
'url': 'http://y.qq.com/#type=singer&mid=001BLpXF2DyJe2',
'info_dict': {
'id': '001BLpXF2DyJe2',
'title': '林俊杰',
'description': 'md5:2a222d89ba4455a3af19940c0481bb78',
},
'playlist_count': 12,
}
def _real_extract(self, url):
mid = self._match_id(url)
singer_page = self._download_webpage(
2015-03-18 17:47:07 +00:00
self.qq_static_url('singer', mid), mid, 'Download singer page')
2015-03-18 06:59:33 +00:00
2015-03-18 17:47:07 +00:00
entries = self.get_entries_from_page(singer_page)
2015-03-18 06:59:33 +00:00
singer_name = self._html_search_regex(
r"singername\s*:\s*'([^']+)'", singer_page, 'singer name',
default=None)
singer_id = self._html_search_regex(
r"singerid\s*:\s*'([0-9]+)'", singer_page, 'singer id',
default=None)
singer_desc = None
if singer_id:
req = compat_urllib_request.Request(
'http://s.plcloud.music.qq.com/fcgi-bin/fcg_get_singer_desc.fcg?utf8=1&outCharset=utf-8&format=xml&singerid=%s' % singer_id)
req.add_header(
'Referer', 'http://s.plcloud.music.qq.com/xhr_proxy_utf8.html')
singer_desc_page = self._download_xml(
2015-03-18 17:47:07 +00:00
req, mid, 'Donwload singer description XML')
2015-03-18 06:59:33 +00:00
singer_desc = singer_desc_page.find('./data/info/desc').text
return self.playlist_result(entries, mid, singer_name, singer_desc)
2015-03-18 17:47:07 +00:00
class QQMusicAlbumIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:album'
2015-03-18 17:47:07 +00:00
_VALID_URL = r'http://y.qq.com/#type=album&mid=(?P<id>[0-9A-Za-z]+)'
_TEST = {
'url': 'http://y.qq.com/#type=album&mid=000gXCTb2AhRR1&play=0',
'info_dict': {
'id': '000gXCTb2AhRR1',
'title': '我们都是这样长大的',
'description': 'md5:d216c55a2d4b3537fe4415b8767d74d6',
},
'playlist_count': 4,
}
def _real_extract(self, url):
mid = self._match_id(url)
album_page = self._download_webpage(
self.qq_static_url('album', mid), mid, 'Download album page')
entries = self.get_entries_from_page(album_page)
album_name = self._html_search_regex(
r"albumname\s*:\s*'([^']+)',", album_page, 'album name',
default=None)
album_detail = self._html_search_regex(
r'<div class="album_detail close_detail">\s*<p>((?:[^<>]+(?:<br />)?)+)</p>',
album_page, 'album details', default=None)
return self.playlist_result(entries, mid, album_name, album_detail)
class QQMusicToplistIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:toplist'
_VALID_URL = r'http://y\.qq\.com/#type=toplist&p=(?P<id>(top|global)_[0-9]+)'
2015-05-14 15:25:43 +00:00
_TESTS = [{
2015-06-04 03:27:18 +00:00
'url': 'http://y.qq.com/#type=toplist&p=global_123',
'info_dict': {
2015-06-04 03:27:18 +00:00
'id': 'global_123',
'title': '美国iTunes榜',
},
'playlist_count': 10,
}, {
2015-06-04 03:27:18 +00:00
'url': 'http://y.qq.com/#type=toplist&p=top_3',
'info_dict': {
2015-06-04 03:27:18 +00:00
'id': 'top_3',
'title': 'QQ音乐巅峰榜·欧美',
2015-06-04 03:27:18 +00:00
'description': 'QQ音乐巅峰榜·欧美根据用户收听行为自动生成集结当下最流行的欧美新歌:更新时间每周四22点|统'
'计周期:一周(上周四至本周三)|统计对象:三个月内发行的欧美歌曲|统计数量100首|统计算法:根据'
'歌曲在一周内的有效播放次数由高到低取前100名同一歌手最多允许5首歌曲同时上榜|有效播放次数:'
'登录用户完整播放一首歌曲记为一次有效播放同一用户收听同一首歌曲每天记录为1次有效播放'
},
'playlist_count': 100,
}, {
2015-06-04 03:27:18 +00:00
'url': 'http://y.qq.com/#type=toplist&p=global_106',
'info_dict': {
2015-06-04 03:27:18 +00:00
'id': 'global_106',
'title': '韩国Mnet榜',
},
'playlist_count': 50,
}]
def _real_extract(self, url):
list_id = self._match_id(url)
2015-05-14 15:28:42 +00:00
list_type, num_id = list_id.split("_")
2015-05-14 15:28:42 +00:00
toplist_json = self._download_json(
2015-06-04 03:27:18 +00:00
'http://i.y.qq.com/v8/fcg-bin/fcg_v8_toplist_cp.fcg?type=%s&topid=%s&format=json'
% (list_type, num_id),
list_id, 'Download toplist page')
2015-06-04 03:27:18 +00:00
entries = [
self.url_result(
'http://y.qq.com/#type=song&mid=' + song['data']['songmid'], 'QQMusic', song['data']['songmid']
) for song in toplist_json['songlist']
]
topinfo = toplist_json.get('topinfo', {})
list_name = topinfo.get('ListName')
list_description = topinfo.get('info')
2015-06-04 03:27:18 +00:00
return self.playlist_result(entries, list_id, list_name, list_description)