Google suddenly stopped indexing my posts after I changed the categories - indexing

Why suddenly Google stopped indexing my posts? I started posting regularly in October. And the site got Adsense approval. I just changed the categories and front design. Some posts were in the top results but they lost ranks too. Domain age 2 yrs.
All of my posts got indexed after I submitted request. But after changing the category structure new posts are not indexing. But when I run live test search console says, "the post is ready to index" and no error sign shows up.
If it’s for the category change then how can I fix that?
I repeatedly submitted imdexing requests after running the live url test. I also submitted a new site map on the search console. But no luck. Google indexed no new post.

Related

Google Merchant Center - Inaccurate availability (due to inconsistent availability between the landing page and checkout pages on your website)

I've gotten a message that my site may be knocked off of Google Merchant Center due to "Inaccurate availability (due to inconsistent availability between the landing page and checkout pages on your website)".
This affects only a small amount of products (only around 0.3% of my 40,000-ish products), so I know it's not an engine issue. After asking Google to recheck the results, they came back with the same error, but with a completely different list of products with no overlap, so I know it's not a problem on the individual product level.
There's no geo-locking on these products, and Google says that the problem exists on US IPs.
Nearly all of the errors look like this:
Value on the landing page - v:out_of_stock
Value in the data feed - v:in_stock
Performing an audit on the products in question shows that none of them have been out of stock for weeks, so the data feed is correct.
None of Google's suggested common issues (geolocking, buy button not working, product can't be shipped to an address, products not available country-wide) seem to apply. The country Google checked this on was a US-based IP.
I'm running out of ideas here, does anyone have any other suggestions?
The answer turned out to be something silly for my site, but I'm posting the answer here just in case this helps someone else.
Google's crawler was setting their country to be Andorra and attempting to check out using the US site. This is obviously not a good representation of the US experience. Google advised us that this was a mistake on their crawler's part, and that we would pass the next audit without any modifications. So if you're here looking for a solution, the absolute best advice I can give you is to find a phone number for Google Merchant Center and give them a call because the error may not be on your end at all.
Update: We passed the audit with no changes made on our part.

spotify API query not producing Album metadata

I'm trying to build a app that gets info from the Spotify API but since I don't have the Spotify ids I want to grab the ids using search queries.
I wanted to just try it so I tested it with an album called yellow and a band called SCANDAL. However when I query, say, https://api.spotify.com/v1/search?q=album:yellow+artist:scandal&type=album it comes up with nothing. However when I Google the band, I got an artist id: 7hTZwqQILVH4bAbN67CeEz from the URL.
using the get albums query(https://api.spotify.com/v1/artists/7hTZwqQILVH4bAbN67CeEz/albums?album_type=album&limit=5) it shows that the album exists but the original query didn't find it. Am I doing something wrong or does it only work with more popular artists (I tried other albums and it was fine)?
There doesn't seem to be anything wrong with your query. I think you are hitting a bug.
Bug reports for Spotify Web API can be found in Github issues
According to this bug:
https://github.com/spotify/web-api/issues/194
setting no market parameter will make it default to US. Your query seems to confirm this bug. Since SCANDAL Yellow is not available in the US, you will not see it in the results. If you add market=SG for instance, it will show up.

Change in number of geotagged Instagram posts

I have written an app which uses the Instagram API to retrieve geotagged posts containing "#sunset" and related tags. On March 22 - 23 2016, the number of posts being collected dropped dramatically. Has anyone else observed this behavior, or do you have any suggestions for what I should look for? Is it possible there has been a change to Instagram's privacy policy that leads to fewer geotagged posts? I do not think it is due to a change in my code (I have not changed the code, and I am actually collecting the posts using two different methods, both of which display this pattern). Nor do I think it is due to my API key suddenly becoming invalid (the key seems to work).
UPDATE: I discovered other people have been having similar problems. See here:
Instagram API /tags/{tag-name}/media/recent changed behaviour
Instagram /v1/tags/{tag-name}/media/recent endpoint doesn't return min_tag_id in pagination block
UPDATE 2: it appears that both the Python + Ruby Instagram clients were deprecated on the same day, March 22, right before everyone started having this problem: https://github.com/facebookarchive/instagram-ruby-gem/commits/master, https://github.com/facebookarchive/python-instagram/commits/master

How to Remove Submitted Content from Search Engines (Google)?

I have submitted 1000 pages to Google included in my sitemap, but I did it by mistake. I didn't want to submit all, I wanted to submit 800 and release 1 page per day during 200 days in order to add new fresh content every day, that way Google would see it as a frequently updated website which is a good SEO practice.
I don't want Google to know about the existence of those 200 pages right now, I want Google to think it is fresh content when I release it every day.
Shall I resend the sitemap.xml with only 800 links and hide the pages in the website?
If Google has already indexed the pages, is there any change to make Google "forget" those pages and not recognize them in the future when I release them back?
Any suggestion about what to do?
Thank you guys.
I wouldn't do that. You're trying to cheat, Google doesn't like this. Remain your site as it is now, create new content and submit it into Google's index as frequently as you want. If you'll exclude previously submitted data from the index, with a high probability it won't be indexed again.

/me/home doesn't return all the posts

I've noticed that me/home in the Graph API doesn't return posts by certain users. I've tried this in my app as well as just using Graph Explorer. It returns most posts, consistently fails to include posts by certain friends. I don't think this is a caching issue, because I've tried over a period of one day with same results. It's not random either. It's the same handful of posts that are always missing.
I checked the posts in question and don't see anything special. And it's happening with newer posts also.
Do I need to add any special parameters to my request ?
Users, in their Facebook privacy preferences, can configure what data 3rd party applications can see about themselves.
What you are seeing is that certain users have made their activities not viewable to your application.