@@ 911-936 (lines=26) @@ | ||
908 | } |
|
909 | ||
910 | ||
911 | class MyExtraLife(GenericNavigableComic): |
|
912 | """Class to retrieve My Extra Life comics.""" |
|
913 | name = 'extralife' |
|
914 | long_name = 'My Extra Life' |
|
915 | url = 'http://www.myextralife.com' |
|
916 | get_navi_link = get_link_rel_next |
|
917 | ||
918 | @classmethod |
|
919 | def get_first_comic_link(cls): |
|
920 | """Get link to first comics.""" |
|
921 | return get_soup_at_url(cls.url).find('a', class_='comic_nav_link first_comic_link') |
|
922 | ||
923 | @classmethod |
|
924 | def get_comic_info(cls, soup, link): |
|
925 | """Get information about a particular comics.""" |
|
926 | title = soup.find("h1", class_="comic_title").string |
|
927 | date_str = soup.find("span", class_="comic_date").string |
|
928 | day = string_to_date(date_str, "%B %d, %Y") |
|
929 | imgs = soup.find_all("img", class_="comic") |
|
930 | assert all(i['alt'] == i['title'] == title for i in imgs) |
|
931 | return { |
|
932 | 'title': title, |
|
933 | 'img': [i['src'] for i in imgs if i["src"]], |
|
934 | 'day': day.day, |
|
935 | 'month': day.month, |
|
936 | 'year': day.year |
|
937 | } |
|
938 | ||
939 | ||
@@ 2274-2298 (lines=25) @@ | ||
2271 | } |
|
2272 | ||
2273 | ||
2274 | class LinsEditions(GenericNavigableComic): |
|
2275 | """Class to retrieve L.I.N.S. Editions comics.""" |
|
2276 | # Also on http://linscomics.tumblr.com |
|
2277 | # Now on https://warandpeas.com |
|
2278 | name = 'lins' |
|
2279 | long_name = 'L.I.N.S. Editions' |
|
2280 | url = 'https://linsedition.com' |
|
2281 | _categories = ('LINS', ) |
|
2282 | get_navi_link = get_link_rel_next |
|
2283 | get_first_comic_link = simulate_first_link |
|
2284 | first_url = 'https://linsedition.com/2011/09/07/l-i-n-s/' |
|
2285 | ||
2286 | @classmethod |
|
2287 | def get_comic_info(cls, soup, link): |
|
2288 | """Get information about a particular comics.""" |
|
2289 | title = soup.find('meta', property='og:title')['content'] |
|
2290 | imgs = soup.find_all('meta', property='og:image') |
|
2291 | date_str = soup.find('meta', property='article:published_time')['content'][:10] |
|
2292 | day = string_to_date(date_str, "%Y-%m-%d") |
|
2293 | return { |
|
2294 | 'title': title, |
|
2295 | 'img': [i['content'] for i in imgs], |
|
2296 | 'month': day.month, |
|
2297 | 'year': day.year, |
|
2298 | 'day': day.day, |
|
2299 | } |
|
2300 | ||
2301 | ||
@@ 341-363 (lines=23) @@ | ||
338 | return [] |
|
339 | ||
340 | ||
341 | class ExtraFabulousComics(GenericNavigableComic): |
|
342 | """Class to retrieve Extra Fabulous Comics.""" |
|
343 | name = 'efc' |
|
344 | long_name = 'Extra Fabulous Comics' |
|
345 | url = 'http://extrafabulouscomics.com' |
|
346 | get_first_comic_link = get_a_navi_navifirst |
|
347 | get_navi_link = get_link_rel_next |
|
348 | ||
349 | @classmethod |
|
350 | def get_comic_info(cls, soup, link): |
|
351 | """Get information about a particular comics.""" |
|
352 | img_src_re = re.compile('^%s/wp-content/uploads/' % cls.url) |
|
353 | imgs = soup.find_all('img', src=img_src_re) |
|
354 | title = soup.find('meta', property='og:title')['content'] |
|
355 | date_str = soup.find('meta', property='article:published_time')['content'][:10] |
|
356 | day = string_to_date(date_str, "%Y-%m-%d") |
|
357 | return { |
|
358 | 'title': title, |
|
359 | 'img': [i['src'] for i in imgs], |
|
360 | 'month': day.month, |
|
361 | 'year': day.year, |
|
362 | 'day': day.day, |
|
363 | 'prefix': title + '-' |
|
364 | } |
|
365 | ||
366 | ||
@@ 367-388 (lines=22) @@ | ||
364 | } |
|
365 | ||
366 | ||
367 | class GenericLeMondeBlog(GenericNavigableComic): |
|
368 | """Generic class to retrieve comics from Le Monde blogs.""" |
|
369 | _categories = ('LEMONDE', 'FRANCAIS') |
|
370 | get_navi_link = get_link_rel_next |
|
371 | get_first_comic_link = simulate_first_link |
|
372 | first_url = NotImplemented |
|
373 | ||
374 | @classmethod |
|
375 | def get_comic_info(cls, soup, link): |
|
376 | """Get information about a particular comics.""" |
|
377 | url2 = soup.find('link', rel='shortlink')['href'] |
|
378 | title = soup.find('meta', property='og:title')['content'] |
|
379 | date_str = soup.find("span", class_="entry-date").string |
|
380 | day = string_to_date(date_str, "%d %B %Y", "fr_FR.utf8") |
|
381 | imgs = soup.find_all('meta', property='og:image') |
|
382 | return { |
|
383 | 'title': title, |
|
384 | 'url2': url2, |
|
385 | 'img': [convert_iri_to_plain_ascii_uri(i['content']) for i in imgs], |
|
386 | 'month': day.month, |
|
387 | 'year': day.year, |
|
388 | 'day': day.day, |
|
389 | } |
|
390 | ||
391 | ||
@@ 1801-1826 (lines=26) @@ | ||
1798 | } |
|
1799 | ||
1800 | ||
1801 | class SafelyEndangered(GenericNavigableComic): |
|
1802 | """Class to retrieve Safely Endangered comics.""" |
|
1803 | # Also on http://tumblr.safelyendangered.com |
|
1804 | name = 'endangered' |
|
1805 | long_name = 'Safely Endangered' |
|
1806 | url = 'http://www.safelyendangered.com' |
|
1807 | get_navi_link = get_link_rel_next |
|
1808 | get_first_comic_link = simulate_first_link |
|
1809 | first_url = 'http://www.safelyendangered.com/comic/ignored/' |
|
1810 | ||
1811 | @classmethod |
|
1812 | def get_comic_info(cls, soup, link): |
|
1813 | """Get information about a particular comics.""" |
|
1814 | title = soup.find('h2', class_='post-title').string |
|
1815 | date_str = soup.find('span', class_='post-date').string |
|
1816 | day = string_to_date(date_str, '%B %d, %Y') |
|
1817 | imgs = soup.find('div', id='comic').find_all('img') |
|
1818 | alt = imgs[0]['alt'] |
|
1819 | assert all(i['alt'] == i['title'] for i in imgs) |
|
1820 | return { |
|
1821 | 'day': day.day, |
|
1822 | 'month': day.month, |
|
1823 | 'year': day.year, |
|
1824 | 'img': [i['src'] for i in imgs], |
|
1825 | 'title': title, |
|
1826 | 'alt': alt, |
|
1827 | } |
|
1828 | ||
1829 | ||
@@ 2545-2573 (lines=29) @@ | ||
2542 | } |
|
2543 | ||
2544 | ||
2545 | class TalesOfAbsurdity(GenericNavigableComic): |
|
2546 | """Class to retrieve Tales Of Absurdity comics.""" |
|
2547 | # Also on http://tapastic.com/series/Tales-Of-Absurdity |
|
2548 | # Also on http://talesofabsurdity.tumblr.com |
|
2549 | name = 'absurdity' |
|
2550 | long_name = 'Tales of Absurdity' |
|
2551 | url = 'http://talesofabsurdity.com' |
|
2552 | _categories = ('ABSURDITY', ) |
|
2553 | get_first_comic_link = get_a_navi_navifirst |
|
2554 | get_navi_link = get_a_navi_comicnavnext_navinext |
|
2555 | ||
2556 | @classmethod |
|
2557 | def get_comic_info(cls, soup, link): |
|
2558 | """Get information about a particular comics.""" |
|
2559 | title = soup.find('h2', class_='post-title').string |
|
2560 | author = soup.find("span", class_="post-author").find("a").string |
|
2561 | date_str = soup.find("span", class_="post-date").string |
|
2562 | day = string_to_date(date_str, "%B %d, %Y") |
|
2563 | imgs = soup.find("div", id="comic").find_all("img") |
|
2564 | assert all(i['alt'] == i['title'] for i in imgs) |
|
2565 | alt = imgs[0]['alt'] if imgs else "" |
|
2566 | return { |
|
2567 | 'img': [i['src'] for i in imgs], |
|
2568 | 'title': title, |
|
2569 | 'alt': alt, |
|
2570 | 'author': author, |
|
2571 | 'day': day.day, |
|
2572 | 'month': day.month, |
|
2573 | 'year': day.year |
|
2574 | } |
|
2575 | ||
2576 | ||
@@ 2483-2511 (lines=29) @@ | ||
2480 | } |
|
2481 | ||
2482 | ||
2483 | class MisterAndMe(GenericNavigableComic): |
|
2484 | """Class to retrieve Mister & Me Comics.""" |
|
2485 | # Also on http://www.gocomics.com/mister-and-me |
|
2486 | # Also on https://tapastic.com/series/Mister-and-Me |
|
2487 | name = 'mister' |
|
2488 | long_name = 'Mister & Me' |
|
2489 | url = 'http://www.mister-and-me.com' |
|
2490 | get_first_comic_link = get_a_comicnavbase_comicnavfirst |
|
2491 | get_navi_link = get_link_rel_next |
|
2492 | ||
2493 | @classmethod |
|
2494 | def get_comic_info(cls, soup, link): |
|
2495 | """Get information about a particular comics.""" |
|
2496 | title = soup.find('h2', class_='post-title').string |
|
2497 | author = soup.find("span", class_="post-author").find("a").string |
|
2498 | date_str = soup.find("span", class_="post-date").string |
|
2499 | day = string_to_date(date_str, "%B %d, %Y") |
|
2500 | imgs = soup.find("div", id="comic").find_all("img") |
|
2501 | assert all(i['alt'] == i['title'] for i in imgs) |
|
2502 | assert len(imgs) <= 1 |
|
2503 | alt = imgs[0]['alt'] if imgs else "" |
|
2504 | return { |
|
2505 | 'img': [i['src'] for i in imgs], |
|
2506 | 'title': title, |
|
2507 | 'alt': alt, |
|
2508 | 'author': author, |
|
2509 | 'day': day.day, |
|
2510 | 'month': day.month, |
|
2511 | 'year': day.year |
|
2512 | } |
|
2513 | ||
2514 | ||
@@ 2769-2795 (lines=27) @@ | ||
2766 | } |
|
2767 | ||
2768 | ||
2769 | class Optipess(GenericNavigableComic): |
|
2770 | """Class to retrieve Optipess comics.""" |
|
2771 | name = 'optipess' |
|
2772 | long_name = 'Optipess' |
|
2773 | url = 'http://www.optipess.com' |
|
2774 | get_first_comic_link = get_a_navi_navifirst |
|
2775 | get_navi_link = get_link_rel_next |
|
2776 | ||
2777 | @classmethod |
|
2778 | def get_comic_info(cls, soup, link): |
|
2779 | """Get information about a particular comics.""" |
|
2780 | title = soup.find('h2', class_='post-title').string |
|
2781 | author = soup.find("span", class_="post-author").find("a").string |
|
2782 | comic = soup.find('div', id='comic') |
|
2783 | imgs = comic.find_all('img') if comic else [] |
|
2784 | alt = imgs[0]['title'] if imgs else "" |
|
2785 | assert all(i['alt'] == i['title'] == alt for i in imgs) |
|
2786 | date_str = soup.find('span', class_='post-date').string |
|
2787 | day = string_to_date(date_str, "%B %d, %Y") |
|
2788 | return { |
|
2789 | 'title': title, |
|
2790 | 'alt': alt, |
|
2791 | 'author': author, |
|
2792 | 'img': [i['src'] for i in imgs], |
|
2793 | 'month': day.month, |
|
2794 | 'year': day.year, |
|
2795 | 'day': day.day, |
|
2796 | } |
|
2797 | ||
2798 | ||
@@ 2515-2541 (lines=27) @@ | ||
2512 | } |
|
2513 | ||
2514 | ||
2515 | class LastPlaceComics(GenericNavigableComic): |
|
2516 | """Class to retrieve Last Place Comics.""" |
|
2517 | name = 'lastplace' |
|
2518 | long_name = 'Last Place Comics' |
|
2519 | url = "http://lastplacecomics.com" |
|
2520 | get_first_comic_link = get_a_comicnavbase_comicnavfirst |
|
2521 | get_navi_link = get_link_rel_next |
|
2522 | ||
2523 | @classmethod |
|
2524 | def get_comic_info(cls, soup, link): |
|
2525 | """Get information about a particular comics.""" |
|
2526 | title = soup.find('h2', class_='post-title').string |
|
2527 | author = soup.find("span", class_="post-author").find("a").string |
|
2528 | date_str = soup.find("span", class_="post-date").string |
|
2529 | day = string_to_date(date_str, "%B %d, %Y") |
|
2530 | imgs = soup.find("div", id="comic").find_all("img") |
|
2531 | assert all(i['alt'] == i['title'] for i in imgs) |
|
2532 | assert len(imgs) <= 1 |
|
2533 | alt = imgs[0]['alt'] if imgs else "" |
|
2534 | return { |
|
2535 | 'img': [i['src'] for i in imgs], |
|
2536 | 'title': title, |
|
2537 | 'alt': alt, |
|
2538 | 'author': author, |
|
2539 | 'day': day.day, |
|
2540 | 'month': day.month, |
|
2541 | 'year': day.year |
|
2542 | } |
|
2543 | ||
2544 | ||
@@ 2404-2430 (lines=27) @@ | ||
2401 | } |
|
2402 | ||
2403 | ||
2404 | class BiterComics(GenericNavigableComic): |
|
2405 | """Class to retrieve Biter Comics.""" |
|
2406 | name = "biter" |
|
2407 | long_name = "Biter Comics" |
|
2408 | url = "http://www.bitercomics.com" |
|
2409 | get_first_comic_link = get_a_navi_navifirst |
|
2410 | get_navi_link = get_link_rel_next |
|
2411 | ||
2412 | @classmethod |
|
2413 | def get_comic_info(cls, soup, link): |
|
2414 | """Get information about a particular comics.""" |
|
2415 | title = soup.find("h1", class_="entry-title").string |
|
2416 | author = soup.find("span", class_="author vcard").find("a").string |
|
2417 | date_str = soup.find("span", class_="entry-date").string |
|
2418 | day = string_to_date(date_str, "%B %d, %Y") |
|
2419 | imgs = soup.find("div", id="comic").find_all("img") |
|
2420 | assert all(i['alt'] == i['title'] for i in imgs) |
|
2421 | assert len(imgs) == 1 |
|
2422 | alt = imgs[0]['alt'] |
|
2423 | return { |
|
2424 | 'img': [i['src'] for i in imgs], |
|
2425 | 'title': title, |
|
2426 | 'alt': alt, |
|
2427 | 'author': author, |
|
2428 | 'day': day.day, |
|
2429 | 'month': day.month, |
|
2430 | 'year': day.year |
|
2431 | } |
|
2432 | ||
2433 | ||
@@ 1953-1979 (lines=27) @@ | ||
1950 | _categories = ('TUNEYTOONS', ) |
|
1951 | ||
1952 | ||
1953 | class CompletelySeriousComics(GenericNavigableComic): |
|
1954 | """Class to retrieve Completely Serious comics.""" |
|
1955 | name = 'completelyserious' |
|
1956 | long_name = 'Completely Serious Comics' |
|
1957 | url = 'http://completelyseriouscomics.com' |
|
1958 | get_first_comic_link = get_a_navi_navifirst |
|
1959 | get_navi_link = get_a_navi_navinext |
|
1960 | ||
1961 | @classmethod |
|
1962 | def get_comic_info(cls, soup, link): |
|
1963 | """Get information about a particular comics.""" |
|
1964 | title = soup.find('h2', class_='post-title').string |
|
1965 | author = soup.find('span', class_='post-author').contents[1].string |
|
1966 | date_str = soup.find('span', class_='post-date').string |
|
1967 | day = string_to_date(date_str, '%B %d, %Y') |
|
1968 | imgs = soup.find('div', class_='comicpane').find_all('img') |
|
1969 | assert imgs |
|
1970 | alt = imgs[0]['title'] |
|
1971 | assert all(i['title'] == i['alt'] == alt for i in imgs) |
|
1972 | return { |
|
1973 | 'month': day.month, |
|
1974 | 'year': day.year, |
|
1975 | 'day': day.day, |
|
1976 | 'img': [i['src'] for i in imgs], |
|
1977 | 'title': title, |
|
1978 | 'alt': alt, |
|
1979 | 'author': author, |
|
1980 | } |
|
1981 | ||
1982 | ||
@@ 2577-2602 (lines=26) @@ | ||
2574 | } |
|
2575 | ||
2576 | ||
2577 | class EndlessOrigami(GenericNavigableComic): |
|
2578 | """Class to retrieve Endless Origami Comics.""" |
|
2579 | name = "origami" |
|
2580 | long_name = "Endless Origami" |
|
2581 | url = "http://endlessorigami.com" |
|
2582 | get_first_comic_link = get_a_navi_navifirst |
|
2583 | get_navi_link = get_link_rel_next |
|
2584 | ||
2585 | @classmethod |
|
2586 | def get_comic_info(cls, soup, link): |
|
2587 | """Get information about a particular comics.""" |
|
2588 | title = soup.find('h2', class_='post-title').string |
|
2589 | author = soup.find("span", class_="post-author").find("a").string |
|
2590 | date_str = soup.find("span", class_="post-date").string |
|
2591 | day = string_to_date(date_str, "%B %d, %Y") |
|
2592 | imgs = soup.find("div", id="comic").find_all("img") |
|
2593 | assert all(i['alt'] == i['title'] for i in imgs) |
|
2594 | alt = imgs[0]['alt'] if imgs else "" |
|
2595 | return { |
|
2596 | 'img': [i['src'] for i in imgs], |
|
2597 | 'title': title, |
|
2598 | 'alt': alt, |
|
2599 | 'author': author, |
|
2600 | 'day': day.day, |
|
2601 | 'month': day.month, |
|
2602 | 'year': day.year |
|
2603 | } |
|
2604 | ||
2605 | ||
@@ 2041-2065 (lines=25) @@ | ||
2038 | } |
|
2039 | ||
2040 | ||
2041 | class ChuckleADuck(GenericNavigableComic): |
|
2042 | """Class to retrieve Chuckle-A-Duck comics.""" |
|
2043 | name = 'chuckleaduck' |
|
2044 | long_name = 'Chuckle-A-duck' |
|
2045 | url = 'http://chuckleaduck.com' |
|
2046 | get_first_comic_link = get_div_navfirst_a |
|
2047 | get_navi_link = get_link_rel_next |
|
2048 | ||
2049 | @classmethod |
|
2050 | def get_comic_info(cls, soup, link): |
|
2051 | """Get information about a particular comics.""" |
|
2052 | date_str = soup.find('span', class_='post-date').string |
|
2053 | day = string_to_date(remove_st_nd_rd_th_from_date(date_str), "%B %d, %Y") |
|
2054 | author = soup.find('span', class_='post-author').string |
|
2055 | div = soup.find('div', id='comic') |
|
2056 | imgs = div.find_all('img') if div else [] |
|
2057 | title = imgs[0]['title'] if imgs else "" |
|
2058 | assert all(i['title'] == i['alt'] == title for i in imgs) |
|
2059 | return { |
|
2060 | 'month': day.month, |
|
2061 | 'year': day.year, |
|
2062 | 'day': day.day, |
|
2063 | 'img': [i['src'] for i in imgs], |
|
2064 | 'title': title, |
|
2065 | 'author': author, |
|
2066 | } |
|
2067 | ||
2068 |