import itertools from .common import InfoExtractor, SearchInfoExtractor from .dailymotion import DailymotionIE from ..utils import smuggle_url, traverse_obj class NetverseBaseIE(InfoExtractor): _ENDPOINTS = { 'watch': 'watchvideo', 'video': 'watchvideo', 'webseries': 'webseries', 'season': 'webseason_videos', } def _call_api(self, slug, endpoint, query={}, season_id='', display_id=None): return self._download_json( f'https://api.netverse.id/medias/api/v2/{self._ENDPOINTS[endpoint]}/{slug}/{season_id}', display_id or slug, query=query) def _get_comments(self, video_id): last_page_number = None for i in itertools.count(1): comment_data = self._download_json( f'https://api.netverse.id/mediadetails/api/v3/videos/comments/{video_id}', video_id, data=b'', fatal=False, query={'page': i}, note=f'Downloading JSON comment metadata page {i}') or {} yield from traverse_obj(comment_data, ('response', 'comments', 'data', ..., { 'id': '_id', 'text': 'comment', 'author_id': 'customer_id', 'author': ('customer', 'name'), 'author_thumbnail': ('customer', 'profile_picture'), })) if not last_page_number: last_page_number = traverse_obj(comment_data, ('response', 'comments', 'last_page')) if i >= (last_page_number or 0): break class NetverseIE(NetverseBaseIE): _VALID_URL = r'https?://(?:\w+\.)?netverse\.id/(?Pwatch|video)/(?P[^/?#&]+)' _TESTS = [{ # Watch video 'url': 'https://www.netverse.id/watch/waktu-indonesia-bercanda-edisi-spesial-lebaran-2016', 'info_dict': { 'id': 'k4yhqUwINAGtmHx3NkL', 'title': 'Waktu Indonesia Bercanda - Edisi Spesial Lebaran 2016', 'ext': 'mp4', 'season': 'Season 2016', 'description': 'md5:d41d8cd98f00b204e9800998ecf8427e', 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'episode_number': 22, 'episode': 'Episode 22', 'uploader_id': 'x2ir3vq', 'age_limit': 0, 'tags': [], 'view_count': int, 'display_id': 'waktu-indonesia-bercanda-edisi-spesial-lebaran-2016', 'duration': 2990, 'upload_date': '20210722', 'timestamp': 1626919804, 'like_count': int, 'uploader': 'Net Prime', }, }, { # series 'url': 'https://www.netverse.id/watch/jadoo-seorang-model', 'info_dict': { 'id': 'x88izwc', 'title': 'Jadoo Seorang Model', 'ext': 'mp4', 'season': 'Season 2', 'description': 'md5:8a74f70812cca267e19ee0635f0af835', 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'episode_number': 2, 'episode': 'Episode 2', 'view_count': int, 'like_count': int, 'display_id': 'jadoo-seorang-model', 'uploader_id': 'x2ir3vq', 'duration': 635, 'timestamp': 1646372927, 'tags': ['PG069497-hellojadooseason2eps2'], 'upload_date': '20220304', 'uploader': 'Net Prime', 'age_limit': 0, }, 'skip': 'video get Geo-blocked for some country', }, { # non www host 'url': 'https://netverse.id/watch/tetangga-baru', 'info_dict': { 'id': 'k4CNGz7V0HJ7vfwZbXy', 'ext': 'mp4', 'title': 'Tetangga Baru', 'season': 'Season 1', 'description': 'md5:23fcf70e97d461d3029d25d59b2ccfb9', 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'episode_number': 1, 'episode': 'Episode 1', 'timestamp': 1624538169, 'view_count': int, 'upload_date': '20210624', 'age_limit': 0, 'uploader_id': 'x2ir3vq', 'like_count': int, 'uploader': 'Net Prime', 'tags': ['PG008534', 'tetangga', 'Baru'], 'display_id': 'tetangga-baru', 'duration': 1406, }, }, { # /video url 'url': 'https://www.netverse.id/video/pg067482-hellojadoo-season1', 'title': 'Namaku Choi Jadoo', 'info_dict': { 'id': 'x887jzz', 'ext': 'mp4', 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'season': 'Season 1', 'episode_number': 1, 'description': 'md5:d4f627b3e7a3f9acdc55f6cdd5ea41d5', 'title': 'Namaku Choi Jadoo', 'episode': 'Episode 1', 'age_limit': 0, 'like_count': int, 'view_count': int, 'tags': ['PG067482', 'PG067482-HelloJadoo-season1'], 'duration': 780, 'display_id': 'pg067482-hellojadoo-season1', 'uploader_id': 'x2ir3vq', 'uploader': 'Net Prime', 'timestamp': 1645764984, 'upload_date': '20220225', }, 'skip': 'This video get Geo-blocked for some country', }, { # video with comments 'url': 'https://netverse.id/video/episode-1-season-2016-ok-food', 'info_dict': { 'id': 'k6hetBPiQMljSxxvAy7', 'ext': 'mp4', 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'display_id': 'episode-1-season-2016-ok-food', 'like_count': int, 'description': '', 'duration': 1471, 'age_limit': 0, 'timestamp': 1642405848, 'episode_number': 1, 'season': 'Season 2016', 'uploader_id': 'x2ir3vq', 'title': 'Episode 1 - Season 2016 - Ok Food', 'upload_date': '20220117', 'tags': [], 'view_count': int, 'episode': 'Episode 1', 'uploader': 'Net Prime', 'comment_count': int, }, 'params': { 'getcomments': True, }, }, { # video with multiple page comment 'url': 'https://netverse.id/video/match-island-eps-1-fix', 'info_dict': { 'id': 'x8aznjc', 'ext': 'mp4', 'like_count': int, 'tags': ['Match-Island', 'Pd00111'], 'display_id': 'match-island-eps-1-fix', 'view_count': int, 'episode': 'Episode 1', 'uploader': 'Net Prime', 'duration': 4070, 'timestamp': 1653068165, 'description': 'md5:e9cf3b480ad18e9c33b999e3494f223f', 'age_limit': 0, 'title': 'Welcome To Match Island', 'upload_date': '20220520', 'episode_number': 1, 'thumbnail': r're:https?://s\d+\.dmcdn\.net/v/[^/]+/x1080', 'uploader_id': 'x2ir3vq', 'season': 'Season 1', 'comment_count': int, }, 'params': { 'getcomments': True, }, }] def _real_extract(self, url): display_id, sites_type = self._match_valid_url(url).group('display_id', 'type') program_json = self._call_api(display_id, sites_type) videos = program_json['response']['videos'] return { '_type': 'url_transparent', 'ie_key': DailymotionIE.ie_key(), 'url': smuggle_url(videos['dailymotion_url'], {'query': {'embedder': 'https://www.netverse.id'}}), 'display_id': display_id, 'title': videos.get('title'), 'season': videos.get('season_name'), 'thumbnail': traverse_obj(videos, ('program_detail', 'thumbnail_image')), 'description': traverse_obj(videos, ('program_detail', 'description')), 'episode_number': videos.get('episode_order'), '__post_extractor': self.extract_comments(display_id), } class NetversePlaylistIE(NetverseBaseIE): _VALID_URL = r'https?://(?:\w+\.)?netverse\.id/(?Pwebseries)/(?P[^/?#&]+)' _TESTS = [{ # multiple season 'url': 'https://netverse.id/webseries/tetangga-masa-gitu', 'info_dict': { 'id': 'tetangga-masa-gitu', 'title': 'Tetangga Masa Gitu', }, 'playlist_count': 519, }, { # single season 'url': 'https://netverse.id/webseries/kelas-internasional', 'info_dict': { 'id': 'kelas-internasional', 'title': 'Kelas Internasional', }, 'playlist_count': 203, }] def parse_playlist(self, json_data, playlist_id): slug_sample = traverse_obj(json_data, ('related', 'data', ..., 'slug'))[0] for season in traverse_obj(json_data, ('seasons', ..., 'id')): playlist_json = self._call_api( slug_sample, 'season', display_id=playlist_id, season_id=season) for current_page in range(playlist_json['response']['season_list']['last_page']): playlist_json = self._call_api(slug_sample, 'season', query={'page': current_page + 1}, season_id=season, display_id=playlist_id) for slug in traverse_obj(playlist_json, ('response', ..., 'data', ..., 'slug')): yield self.url_result(f'https://www.netverse.id/video/{slug}', NetverseIE) def _real_extract(self, url): playlist_id, sites_type = self._match_valid_url(url).group('display_id', 'type') playlist_data = self._call_api(playlist_id, sites_type) return self.playlist_result( self.parse_playlist(playlist_data['response'], playlist_id), traverse_obj(playlist_data, ('response', 'webseries_info', 'slug')), traverse_obj(playlist_data, ('response', 'webseries_info', 'title'))) class NetverseSearchIE(SearchInfoExtractor): _SEARCH_KEY = 'netsearch' _TESTS = [{ 'url': 'netsearch10:tetangga', 'info_dict': { 'id': 'tetangga', 'title': 'tetangga', }, 'playlist_count': 10, }] def _search_results(self, query): last_page = None for i in itertools.count(1): search_data = self._download_json( 'https://api.netverse.id/search/elastic/search', query, query={'q': query, 'page': i}, note=f'Downloading page {i}') videos = traverse_obj(search_data, ('response', 'data', ...)) for video in videos: yield self.url_result(f'https://netverse.id/video/{video["slug"]}', NetverseIE) last_page = last_page or traverse_obj(search_data, ('response', 'lastpage')) if not videos or i >= (last_page or 0): break