Like Ra in latex catsuit, latex mask and high heels
Like Ra's Naughty Playground

shapewear
J0PF Women Wide Laceup Corset Slimming Waist Belt Shapewear Stage Dance Performances Laceup Waistbelt Corset
$261.44

"inlzdz"
Fringe Latin Dance Dress for Women Latin Outfit Ladies Flowy Tassel Competition Dress Salsa Tango Samba Rumba Dance Costume
$14.86-40%

bondage device
Male Chastity Device Flat Round Hole Metal Chastity Lock,Cock Cage,Urethra Catheter,Penis Ring,Bondage Belt Sex Toys For Men Gay
$23.96-53%

hentai anime adult
NSFW Nude 3d Nipples Hentai Boobs Mouse Pad Sono Bisque Doll Wa Koi Wo Suru Kitagawa Marin Anime Wrist Rest 3D Oppai MousepadCos
$186.17-60%

erotic
Vtg Slayyyter Erotic Electronic Cotton Black Full Size Unisex Shirt
$108.90-41%

"crossdress"
CD Silicone Adhesive Roll-On Emulsion Binder Chest Stickers Invisible Fake Silicone Breast Glue Water Crossdresser Skin Special
$24.18-6%

"sissy" "femboy"
Femboy Maid Anime Boy Aesthetic Crossdressing T Shirt Big Size 100% Cotton Femboy Anime Boy Maid Neko Sissy Queer Mtf Bisexual
$93.17-5%



To view Aliexpress you might need to switch your mobile browser to the Desktop version.


House of self-bondage
House of self-bondage
From €450 per week

If you would like to use search functions, view hidden archives or play games, please consider registering.

No activation mail? (Gmail is known for that) Please contact me directly!

We do not have old threads! Feel free to post in any! Regardless the age!


Imgur will no longer host NSFW content as of May 15
#1
The mods at r/hentai have made a better write up than I could, I suggest you read it here.

TL;DR: Imgur to prohibit NSFW content, remove all previously uploaded NSFW content, and to delete ALL content uploaded by unregistered users.
Reply
#2
I should mention, if you are an avid user of Reddit, and have images and video saved, most likely it was uploaded via imgur, so you will want to grab local copies of them before your saves are filled with dead links
Reply
#3
See why I insist on uploading all media to the forum?
See why we need money for storage? https://www.likera.com/forum/mybb/Thread...g-payments

... hint-hint ... wink-wink ...
Reply
#4
Well, back to Tumblr.

🎵 It's the ciiiircle of liiiiiife 🎶

What will they do with all the petabytes of free disk space, when no one uses their service anymore?

Removing pornography I would understand (and support even, in a way), but nudity?
Lol, good luck.
Life's too short to pretend you're not into some kinky shit.
Reply
#5
(21 Apr 2023, 18:00 )Bound Whore Wrote: Well, back to Tumblr.
Tumblr was the first one to ban and delete adult images.
Reply
#6
For those of you who have upvoted (or saved) quite a bunch of posts with Imgur-hosted images over on reddit.com and know their way around Python, there's a script that allows you to download images that are embedded within posts. Amongst others it will download images from Imgur (it also download e.g. images directly embedded into posts).

I took the script from here and made some changes because I had problems with the original: https://pastebin.com/t5E5sTY2

By switching out current_redditor.saved with current_redditor.upvoted on line 160, you can switch between saved and upvoted posts.

I'll also put the code below. I don't expect pastebin to shut down anytime soon, but better save then sorry.

Code:
import requests
from datetime import datetime
import os
import pandas as pd
import praw.models
import urllib.request
import logging, sys
from bs4 import *

# How to use:
# 1. Install python3 - https://www.python.org/downloads/
# 2. Open "Command prompt" on your PC and copy and paste: `pip install pandas, bs4, urllib` (Without quotes)
# 3. Fill in details below (Link explains how)
# 4. Run it this file, and it will download your last 1000 upvoted posts (1000 is the max set by reddit)

# Fails to work for: Redgifs, bdsmlr
# Downloads crap along with correct post for: gfycat (Also fails half of the time)

# Fill in your details here
# https://praw.readthedocs.io/en/stable/getting_started/authentication.html#password-flow
reddit = praw.Reddit(
    client_id="",
    client_secret="",
    password="",
    user_agent="Downloads images from /u/<username>/upvoted before Imgur deletes them all",
    username=""
)

column_list = ["title", "post_url", "user", "image_url", "image_loc", "notes"]
upvoted_df = pd.DataFrame(data=None,
                          index=None,
                          columns=column_list,
                          dtype=None,
                          copy=None
                          )


def clean_title(submission_title: str) -> str:
    """
    Remove all values not allowed in Windows file names
    Makes name shorter than max file length Windows allows
    :param submission_title:
    :return:
    """
    for bad_char in list('\\/:*?\"<>|'):
        submission_title = submission_title.replace(bad_char, "#")
    return submission_title[:180]


def download_images(url, folder_name) -> None:
    """
    Download all images from URL
    From:
    https://www.geeksforgeeks.org/how-to-download-all-images-from-a-web-page-in-python/

    #TODO Doens't work with redgifs
    :param url: URL to download all images from
    :param folder_name: Relative folder destination for images
    :return:
    """
    # content of URL
    r = requests.get(url)

    # Parse HTML Code
    soup = BeautifulSoup(r.text, 'html.parser')

    for thing in ["img", "video"]:
        # find all images in URL
        images = soup.findAll(thing, limit=100)

        # initial count is zero
        count = 0

        # print total images found in URL
        print(f"Total {len(images)} {thing} Found!")

        # checking if images is not zero
        if len(images) != 0:
            for i, image in enumerate(images):
                # From image tag ,Fetch image Source URL

                # 1.data-srcset
                # 2.data-src
                # 3.data-fallback-src
                # 4.src

                # Here we will use exception handling

                # first we will search for "data-srcset" in img tag
                try:
                    # In image tag ,searching for "data-srcset"
                    image_link = image["data-srcset"]

                # then we will search for "data-src" in img
                # tag and so on
                except:
                    try:
                        # In image tag ,searching for "data-src"
                        image_link = image["data-src"]
                    except:
                        try:
                            # In image tag ,searching for "data-fallback-src"
                            image_link = image["data-fallback-src"]
                        except:
                            try:
                                # In image tag ,searching for "src"
                                image_link = image["src"]

                            # if no Source URL found
                            except:
                                pass

                # After getting Image Source URL
                # We will try to get the content of image
                try:
                    # Warning is fine as it is in try/except
                    r = requests.get(image_link).content
                    try:

                        # possibility of decode
                        r = str(r, 'utf-8')

                    except UnicodeDecodeError:

                        # After checking above condition, Image Download start
                        with open(f"{folder_name}/image{i + 1}.jpg", "wb+") as f:
                            f.write(r)

                        # counting number of image downloaded
                        count += 1
                except:
                    print(f"Could not find content for '{image_link}' ({requests.get(image_link)})")
                    pass

            # There might be possible, that all
            # images not download
            # if all images download
            if count == len(images):
                print(f"All {thing}s Downloaded!")

            # if all images not download
            else:
                print(f"Total {count} {thing}s downloaded out of {len(images)}")


print(f"Downloading upvoted posts for: {reddit.user.me()}")
current_redditor: praw.models.Redditor = reddit.user.me()
cwd = os.path.dirname(__file__)

try:
    os.mkdir(f"{cwd}/images/")
except FileExistsError:
    print("/images/ already exists")
try:
    os.mkdir(f"{cwd}/posts/")
except FileExistsError:
    print(f"/posts/ already exists")

# Max limit that PRAW allows easily (1000)
for counter, submission in enumerate(current_redditor.saved(limit=1000)):
    try:
        submission: praw.models.Submission  # So editor knows

        filetype: str
        to_append: pd.DataFrame

        if not hasattr(submission, 'title'):
            print("Found a comment.")
            print(submission.link_title)
            print(submission.link_permalink)
            print("Skipping.")
            continue

        title = clean_title(submission.title)
        title_with_counter = f"{counter}-{title}"
        author = submission.author
        if author is None:
            author = "[deleted]"
        else:
            author = submission.author.name

        # If a url link
        if submission.selftext == "":
            # If image/video link
            # https://help.imgur.com/hc/en-us/articles/115000083326-What-files-can-I-upload-Is-there-a-size-limit-
            (_, filetype) = os.path.splitext(submission.url)
            if filetype.upper() in [".PNG", ".GIF", ".JPG", ".JPEG", ".MP4", ".MPEG", ".AVI", ".WEBM", ".APNG", ".TIFF",
                                    ".MOV", ".QT", ".MKV",
                                    ".MK3D", ".MKA", ".MKS", ".FLV", ".F4V", ".F4P", ".F4A", ".F4B"]:
                print(f"Directly Downloading: '{submission.url}' as {filetype}")

                image_loc = f"{cwd}/images/{title_with_counter}.{filetype}"

                # Save image
                urllib.request.urlretrieve(submission.url, image_loc)

                df_row = pd.DataFrame(
                    [
                        [
                            submission.title,
                            submission.permalink,
                            author,
                            submission.url,
                            image_loc,
                            "IMAGE"
                        ]
                    ],
                    columns=column_list)

            # Non-Image url
            # Download all images on page
            else:
                print(f"Downloading files on page for: '{submission.url}'")

                image_folder_loc = f"{cwd}/images/{title_with_counter}/"
                try:
                    os.mkdir(image_folder_loc)
                except FileExistsError:
                    print(f"/images/{title_with_counter} already exists")

                download_images(submission.url, image_folder_loc)

                df_row = pd.DataFrame(
                    [
                        [
                            submission.title,
                            submission.permalink,
                            author,
                            submission.url,
                            image_folder_loc,
                            "IMAGE FOLDER"
                        ]
                    ],
                    columns=column_list)

        # If non-url (text) post
        # TODO could be Poll I guess
        else:
            print(f"Downloading Text For: '{submission.url}'")
            txt_loc = f"{cwd}/posts/{counter}-{title}.txt"
            with open(txt_loc, "w+") as file:
                file.write(submission.selftext)

            df_row = pd.DataFrame(
                [
                    [
                        submission.title,
                        submission.permalink,
                        author,
                        "",
                        txt_loc,
                        "TEXT POST"
                    ]
                ],
                columns=column_list)

        # Append to df
        upvoted_df = pd.concat([upvoted_df, df_row])

    except Exception:
        print(f"Failed to download {submission.title}")
        df_row = pd.DataFrame(
            [
                [
                    submission.title,
                    submission.permalink,
                    "FAILED",
                    "",
                    "FAILED",
                    "FAILED"
                ]
            ],
            columns=column_list)
        upvoted_df = pd.concat([upvoted_df, df_row])

upvoted_df.to_csv(f"{str(datetime.now()).replace(':', '-')}.csv")

Reply
#7
Can't gallery-dl be used to download from imgur?

https://github.com/mikf/gallery-dl
Reply




Contributors: Bound Whore (1) , Like Ra (3) , Thegolor (1) , TiedTight (2)