The google search engine is an art and a science. To begin with google trawls and indexes trillions of documents and pages a day. By using software called “web crawlers” they find publicly accessible webpages. By documenting these pages makes it accessible to everyone else not he internet. Websites owners can set up restrictions for “crawling”, changing what is shown as a summary of the page below the title or a cached version in case the live version is unavailable.
Spam is a major problem for search engines. Spam sites use tricks like invisible text and repeating a word to get their pages to the top of the search results. Googles algorithms can detect spam sites the majority of the time and put them at the bottom of a search but for the other occasions google teams review the sites.
Search engine optimisation is a technique employed by website managers to improve their occurrence when searched for. Manipulating the content of a website can cause it to be pushed to the top of a search so it is viewed by more people than if it was at the bottom of the search. the key “Ingredients” for SEO are titles, words throughout the page, links and the frequency of visitors to the site, all these aspects are considered when indexing a website.
There are also policies that have to be put in place to ensure certain information such as credit card numbers etc… remains private and that damaging content such as malware is removed. Access to information is a tricky situation free expression is a high concern with google but they must also manage things like malware legal requirements and preventing identity theft. In most cases the use of googlebots instead of manual labour is an effective way organise the search results. The reason being is algorithms are easily calibrated so improvements can be made fairly easily changing the requirements in the algorithms affects every search evenly. Manual exceptions have to be made to some sites as the algorithms cannot interpret the data given to it and can wrongly identify a webpage.
When google’s teams assess a page to be spam related they do their best to notify the web owner of any discrepancies and allow them to file a reconsideration request if they have modified the website to an improved standard. Google will remove certain information from search results if after a request they feel that you might be susceptible to identity theft and fraud. They will however deny a request for information removal if it’s seen as trying to abuse this policy for other reasons.
Legal removals occur when a breach of the law occurs. These laws can come from the Digital Millennium Copyright Act (DMCA) or from versions consistent with local law. When content is removed a notification of this occurrence is left in place of the search result. Content that is potentially disturbing is blocked from the search when not specifically searched for. Some search features are also disabled for queries where the results could be offensive. Google safe search is an effective search mechanism that when used blocks all content that could be pleasant to the viewer. With these policies in place google is taking steps towards ensuring that the content viewed on the internet is specific to the search.