Major search engines redirect an HTTP link to a scammy pharmacy site

There is an HTTP website (not HTTPS) that works perfectly fine when the URL is directly typed into the address bar or when links are clicked from other websites/applications like Reddit, Facebook, and Discord.

The exceptions are the major search engines: Google, Yahoo, and Bing. When clicked from one of these sites, it is redirected to a scammy pharmacy site with one of many different domain names. This occurs in Chrome, Firefox, and Edge; it also occurs on Android smartphones. (Bing only does this in Chrome; it works fine in Firefox and Edge.) This issue occurs for multiple people on many different devices.

Interestingly, other search engines (Dogpile, Baidu, Ask, DuckDuckGo, Yandex, etc.) seem to work fine.

What could be the cause of this behavior? Do the search engines or the website need to be fixed, and how? Would converting to HTTPS help, and why?

The website in question is I am not the webmaster, but as a programmer and web developer I am curious about this issue.

[]⭐Premium Real-Time HIGH LPM GSA SER Link lists⭐★★10% Discount => $22.49★★

GSA SER Site List | GSA SER Verified lists by

❤️❤️Drop a Message and We will send you an exclusive 10% coupon❤️❤️


⚡⚡⚡⚡High LPM⚡⚡⚡⚡
Updating in Real-Time Using Dropbox (set once and forget)

Dedupe Domains (Not URLs!) each 5 Minutes

Contextual & Low OBL Targets

.Edu & .Gov Domains Included

GSA Captcha Breaker Compatible

Over 4M URLs Scraped each Day

Full Automated Scraping and Processing system. Don’t Worry about Real-Time Updates anymore!

We just Scrape Google

Identifying around 1,000,000 Links Each Day

What do we use for making our link lists?

We Got Awesome AMD Ryzen9 Dedicated Servers

GSA SER + GSA CB + GSA PI + Scrapebox

Catchall emails

50 Dedicated Proxies(Processing)

Unlimited Bandwidth Backconnect Proxy(Scraping)

⚡⚡⚡ Place your Order at ⚡⚡⚡

Secure Payments Powered by WarriorPlus


  • All Scraped URLs
  • Contextual Targets
  • Low OBL(outbound links<100)
  • Submitted Targets
  • Verified Targets
Submitted Targets Stats (4/21/2021)

Real Testimonials from our clients

We are Identifying so fast

UPDATES are Real-Time


⚡If you are using our site lists, to get higher LPM & VPM, use “All Scraped Urls” and “Submitted Targets” Folders⚡

Redirects in URL — link juice or not?

Guys, I’m wondering. 
If you have a URL with a redirect, e.g.

or the (old) phpinfo exploits, how could that possibly make google count this as a link?
And if I add a tier to it, would be getting the link juice, not
Or am I wrong? 
I am seriously confused.

Source link plugin – show just anchor / link

I implemented a small source link plugin on my WordPress website via the following codes.


                        <?php global $  post, $  pages, $  page;  $  total = count( $  pages ); // Link źródłowy if ( $  total < 2 || $  page === $  total ) :          if ( $  url = get_post_meta( $  post->ID, '_source_link', true ) ) :         $  label = get_post_meta( $  post->ID, '_source_link_label', true );         $  label = $  label ? $  label : $  url;     ?>         <div class="source-link">             <b>Źródło:</b> <a href="<?php echo esc_url( $  url ); ?>" rel="nofollow" target="_blank"><?php                 echo esc_html( $  label ); ?></a>         </div>     <?php endif;  endif;  ?> 


add_action( 'add_meta_boxes', 'wpse_source_link' );   add_action( 'save_post', 'wpse_source_link_save' );   function wpse_source_link() {      add_meta_box(         'source_link',         __( 'Link źródłowy', 'myplugin_textdomain' ),          'wpse_source_meta_box',         'post',         'side'     ); }   function wpse_source_meta_box( $  post ) {      wp_nonce_field( plugin_basename( __FILE__ ), 'myplugin_noncename' );       echo '<label for="source-link">Link</label> ';   echo '<input type="text" id="source-link"" name="source_link" value="'.     get_post_meta( $  post->ID, '_source_link', true ) .'" size="25" />';    echo '<label for="source-link-label">Nazwa strony</label> ';   echo '<input type="text" id="source-link-label"" name="source_link_label" value="'.     get_post_meta( $  post->ID, '_source_link_label', true ) .'" size="25" />'; }   function wpse_source_link_save( $  post_id ) {    if ( defined( 'DOING_AUTOSAVE' ) && DOING_AUTOSAVE )        return;    if ( ! wp_verify_nonce( $  _POST['myplugin_noncename'], plugin_basename( __FILE__ ) ) )       return;     if ( current_user_can( 'edit_post', $  post_id ) ) {        update_post_meta( $  post_id, '_source_link', sanitize_text_field( $  _POST['source_link'] ) );       update_post_meta( $  post_id, '_source_link_label', sanitize_text_field( $  _POST['source_link_label'] ) );     } } 

As you can see there are two fields: Link źródłowy (Source link) and Nazwa strony (website name – anchor). However, at this moment this plugin only works, when both fields have some text inside. Is there any way to make it work also with just anchor/link? I mean, if two fields are completed it should show anchor with link, but if just anchor has some text in it, then it should show only anchor. Same goes for link.

Does anyone know how to make it work like that? To be honest I’m a newbie and I have no idea.

Fetching content binary from database or fetching content by its link from storage service

For an app (web + phone) there are two options:

  1. Image binaries in database. Server replies to app HTTP request with images as base64
  2. Images in storage service like Amazon S3 or Azure Blob Storage or a self-hosted one. Image links in database. Server handles app HTTP requests by sending back only the links to images. The app fetches the images from storage by their link

2 options

Which option above is the standard practice? Which one has less trouble down the road?

Poisson GeneralizedLinearModelFit with log link, first argument not a vector, matrix or list


I’m trying to plot a graph for an exponential decay for a radioactivity experiment, so model is y=Ae^(-kt). Cross-Validated SE, suggested trying a Poisson GLM with a log link fit, and then to output mean deviance for goodness of fit. I’ve tried doing this but am getting the error:

GeneralizedLinearModelFit::notdata: The first argument is not a vector,  matrix, or a list containing a design matrix and response vector. 

And admittedly, I don’t really understand the documentation for it. And am unsure how to implement the show mean residuals part of this. The documentation shows how to output all of the deviance residuals, but not the mean of them directly.


dataHist5 = {{Around[16.5, 1.5],     Around[77.8, 8.8]}, {Around[34.5, 1.5],     Around[60.5, 8.0]}, {Around[52.5, 1.5],     Around[63.8, 8.0]}, {Around[106.5, 1.5],     Around[42.4, 6.5]}, {Around[124.5, 1.5],     Around[41.7, 6.5]}, {Around[142.5, 1.5],     Around[14.6, 3.8]}, {Around[160.5, 1.5],     Around[33.9, 5.8]}, {Around[178.5, 1.5],     Around[29.4, 5.4]}, {Around[196.5, 1.5],     Around[33.5, 5.8]}, {Around[214.5, 1.5],     Around[30.9, 5.6]}, {Around[232.5, 1.5],     Around[31.1, 5.8]}, {Around[250.5, 1.5],     Around[21.5, 4.6]}, {Around[268.5, 1.5],     Around[4.3, 2.1]}, {Around[286.5, 1.5],     Around[6.4, 2.5]}, {Around[322.5, 1.5],     Around[7.5, 2.7]}, {Around[340.5, 1.5],     Around[4.5, 2.1]}, {Around[358.5, 1.5],     Around[11., 3.3]}, {Around[376.5, 1.5],     Around[14.0, 3.7]}, {Around[394.5, 1.5],     Around[14.0, 3.7]}, {Around[466.5, 1.5],     Around[0.6, 0.7]}, {Around[502.5, 1.5],     Around[2.2, 1.5]}, {Around[520.5, 1.5],     Around[9.4, 3.1]}, {Around[538.5, 1.5],     Around[4.1, 2.0]}, {Around[646.5, 1.5],     Around[2.2, 1.5]}, {Around[682.5, 1.5], Around[0.6, 0.7]}} 

Code, so far:

glm = GeneralizedLinearModelFit[dataHist5, x, x,     ExponentialFamily -> "Poisson"] // Normal Show[ListPlot[dataHist5, Plot[glm[x]]] 

I think I’m missing an argument in Plot[glm[x]]], but not sure what.

Link Extractor saved files are blank

This is really weird because I just scraped like 1800 URLs that found 20X that number of internal links. All of the links are in the TXT file

Then I try to scrape ONE URL,, and it finds 70 internal links but when I open the TXT file, there’s nothing there.

I also disabled my proxies (they were getting a “socket error # 10054”) and then I tried to scrape the homepage of The Onion as a test.

Same problem, it says that 33 internal links were found but nothing saved to the file.

And I tried using my three ProxyMesh proxies, my StormProxies that I just bought today, and no proxies. Every time it says it’s saved but the file is empty.

Any ideas?

Also, is there a way to save scraped data with a more descriptive name? It’s really difficult to find files when the file name is just a long string of numbers.