cache – What do I do if my WordPress site is suddenly sanitizing portions of URLs it shouldn’t?

Nobody has touched wp-admin for this site in weeks. Suddenly, URLs are being sanitized in ways that break everything. The question marks of query parameters are being replaced with URL escape code %3F, which is obviously breaking nearly every script and stylesheet include that’s affected. The result is content like this:

<script type='text/javascript' src=""></script>

<script type='text/javascript' src='wp-includes/js/wp-embed.min.js%3Fver=5.4.1'></script>

Some, but not all URLs in the document are affected. Scripts and some anchor elements are affected.

In addition to this URL madness, it seems like portions of the main site are being included in every single page – i.e., the homepage’s markup is being embedded within wp-admin pages, etc.

The effect is that the site can essentially not be used – large dynamic portions of pages, especially the main landing page, simply fail to load or appear malformed due to missing scripts and stylesheets.

I should add that it’s nearly impossible to navigate wp-admin given the absurd state that it’s in at the moment, so I can’t really find my way around to do regular diagnostics.

How do I even begin to fix something like this? If this wasn’t a moneymaking site with plenty of content I haven’t even touched (‘the last guy’ decided to install dozens of plugins whose purposes are difficult to decipher, are possibly redundant, and might break everything if removed) I would gladly just reinstall everything.

I have attempted:

  • Removing a “WP Fastest Cache” plugin which, according to some, ‘breaks everything’ as of this month. Edit: It looks like this plugin is parasitic. After deactivating it, clearing my local cache, etc., pages still contain a comment marking them as having been cached by it…

  • Visiting and re-saving permalink settings in wp-admin as suggested by a comment I found in a similar issue elsewhere.

How do I get the full URL from a list of short URLs?

Hello everyone. Does anyone know of an online site or anything that I can use to get the full URL of a list of shortened URLs?

I don't want to visit them all individually, that would take too long.

I'm talking about a list of shortened URLs that is about 500 URLs long.

So it would take too long for them to get the full URL for them.

I need to know an online site / tool that I can use to paste the URLs and click Submit. This basically gives me the full URL for them.

Does anyone know anything like that?

Thanks, people!


Block and shorten URLs EARN MONEY – URL shorteners with CPA – Other ways to earn money is a URL shortening service in combination with content blocking. The sales model of this network is CPA. So you won't be paid for clicks and impressions. Users need to take action such as: B. install software or fill out a simple form.

How do I start with

If you shorten the link with, it will be locked automatically. The visitor has to do a simple task to unlock the link. You will receive a commission of $ 0.10 to $ 10 for each task.


You will be paid within 2 working days once you have reached the minimum payment. The payment option is only Bitcoin. The minimum payout is $ 20.

Referral program

The publishers can earn extra money by referring others to this network. The commission rate is 10% for life.

Best properties

-This network accepts members worldwide.

– You do not need a website or blog to use this service.

-Tools (website widget, browser button and API)

– General system

-The members always receive a timely payment.

-Dedicated support

For tips and ideas:

Ref Link:


8 – Is the table path_alias queried for all URLs?

I am trying to generate SEO friendly URLs for a number of exposed filters. For example:

/analyst-relations/firm?related_firm=5072 would look like


One way I imagined was to programmatically add / update / delete this path to the path_alias table when creating / changing / deleting the relevant term or entity. However, I seem to have misunderstood how path_alias works. I was able to add the path successfully /analyst-relations/firm?related_firm=5072 and the alias /analyst-relations/firm/forrester, but the exposed filter connections are still loaded like the former. Isn't the path_alias table queried for all URLs? If not, is it possible to somehow append this path and alias to an entity?

Nginx configuration: Specify a different location for the fallback URLs in a separate file

The standard Nginx configuration loads everyone *.conf Files from /etc/nginx/conf.d Directory. They are then processed in alphabetical order, so I added two files in this directory: 000-default.conf and 001-test.conf. All other conf file names have the format hostname.conf.

The 000-default.conf looks something like this:

server {
    listen       80;
    server_name  localhost;

    access_log  /var/log/nginx/access.log  main;

    location / {
        root   /usr/share/nginx/html;
        index  index.html index.htm;

The 001-test.conf looks something like this:

server {
    listen       80;
    server_name  localhost-test;

    access_log  /var/log/nginx/access.log  main;

    location /test {
        alias   /usr/share/nginx/test;
        index  index.html index.htm;

I would like that / and /test Requests to be delivered from the appropriate sites if the host does not match any of the supported domains configured in the configuration hostname.conf Files. I was forced to change that server_name localhost; to server_name localhostx; by doing 001-test.conf because otherwise there would be the server name conflict and the 001-test.conf would be ignored. But that also means that the 001-test.conf The configuration is only used when I send a request to the server localhostx Domain. So I can only achieve what I want if I move it location /test { Block to 000-default.conf. But is there a way that I can keep? /test Configuration in a separate file? But I can do it include Directive, but then the 001-test.conf must be on a different path or will it be included twice? And then it would be pretty naked, just that location Directive. I'm looking for the best practice proposal here.

Include URLs in blog comments

Does anyone know how to properly insert URLs (with anchors) in blog comments?
(Mine?) GSA SER seems to fail a lot, although all pages are full of comments with URLs with anchors inserted. It can make a difference for my projects.
I tried with:
Enables the "Always try to insert a link in comments" option.
Nothing works and I'm fed up …
Much work in vain, preparation of the lists and so on. Just when I started, I had to stop.
And the number of (empty) anchors makes the difference. I am not sure why GSA SER cannot insert these links.

Python – Use Pymongo to use multiprocessing to scratch over 4 million URLs and investigate the effects of the corona virus

I would therefore like to do some research on the effects of covid 19 on companies. I managed to create a database with the company name and associated website URLs. Now I want to scrape them all off as quickly as possible so I can analyze them. I am new to parallel programming and am skeptical that I will connect to every database as securely as possible.

from __future__ import division

from multiprocessing import Pool

import pymongo as pym
import requests
from bs4 import BeautifulSoup

# Set up local client
client = pym.MongoClient('mongodb://localhost:27017/')
# Connect to local DB
db = client.local_db
# Connect to Collections
My_Collection = db.MyCollection
ScrapedPagesAprilCollection = db.ScrapedPagesApril

# I don't want to scrape these

def parse(url):
    if any(domain in url for domain in LIST_OF_DOMAINS_TO_IGNORE):
    elif '.pdf' in url:
        # print(url)
        headers = {
            'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0',
        page = requests.get(url, headers=headers)
        print(f'{url}: {page.status_code}')
        if page.status_code == 200:
            soup = BeautifulSoup(page.text, 'lxml')
            text = soup.get_text(separator=" ")
            info_to_store = {
                '_id': url,
                'content': text

            if 'coronavirus' in text:
                info_to_store('Impacted') = True

            # Insert into Collection
                {'_id': url}, info_to_store, upsert=True)

        elif page.status_code != 200:
            print(f'{url}: {str(page.status_code)}')

def covid19_scrape_pages(collection, query: dict):
    Wanting to update the pages already matched

    collection : pymongo.collection.Collection
    query : dict

    A url

    # Get the cursor
    mongo_cursor = collection.find(query, no_cursor_timeout=True)
    # For company in the cursor, yield the urls
    for company in mongo_cursor:
        for url in company('URLs'):
            doc = ScrapedPagesAprilCollection.find_one({'_id': url})
            # If I haven't already scraped it, then yield the url
            if doc is None:
                yield (url)

def main():
    print('Make sure LIST_OF_DOMAINS_TO_IGNORE is updated by running',
          ' first')
    urls_gen = covid19_scrape_pages(
        My_Collection, {})
    pool = Pool(8), urls_gen)

if __name__ == "__main__":  # Required logic expression

magento2.3 – Nginx: How do I redirect requests for deleting .html suffixes to the new URLs?

Today I deactivated the suffixes of my Magento installation and deleted older entries with double suffixes (e.g. .html.html) from the database.

To keep existing hyperlinks and search engine entries, I want Nginx to redirect all static page requests with the suffix .html or .html.html to the new path.

Inquiries to

should be redirected to

How do I achieve my goal and what rewriting rules do I need to add to the Nginx configuration?

Save URLs as .php files?

I am transferring my website from another CMS to WordPress. Actually the URLs are as .php files. For example,
However, WP saves and shows URLs as
Is there any way to have urls like before in WP like .php files?
The reason is to keep the site structure and URLs in order to make as few changes as possible to avoid lowering the ranking.
Many thanks