Content of the material
What is a robots.txt file
If you don’t want crawlers to access sections of your site, you can create a robots.txt file with appropriate rules. A robots.txt file is a simple text file containing rules about which crawlers may access which parts of a site. For example, the robots.txt file for example.com may look like this:
How to compress a PDF file in Google Drive
Data compression means reducing the size of a file. However, while transmitting data, this process is called source coding as it means encoding done at the source before sending or storing data.
Steps to follow
Start by choosing the file you want to compress. It can be either from your cloud storage like DropBox, Google Drive, or your computer.
- Automatic Size Reduction. The uploaded file reduces in size automatically once it gets uploaded to the Google Drive system. It keeps a relevant compressed quality suitable for the internet.
- There is the option to further shrink your file to web/email quality.
- When the file is ready, you can access the compressed PDF file by downloading it and viewing it in your Computer’s web browser.
Facts About File Compression
PDF file compression reduces bits of the original file by encoding slightly fewer information. However, there is no information loss. The compressed PDF files target and eliminate any statistical redundancies.
Other PDF Compression Options
There are other advanced compressing options than Google Drive. You can download and install desktop options or compress PDF files using online software.
When compressing PDF files using Google Drive, you have the option to store them and preview them online or offline.
To view a PDF file offline, click on the offline tab and download for a later preview.
Reasons to compress files to PDF
If you can, avoid sharing files in word or PowerPoint form. If there are no reasons why the file should remain as it is, always convert to PDF.
There are two primary reasons why a file should be stored and shared as a PDF file.
When you share information with the public, they can use it elsewhere. Not everyone uses Microsoft products like Excel, PowerPoint, or word since you must purchase them as Microsoft Office products.
New versions of Microsoft products keep changing their file extensions. Different kinds of file extensions like pptx and docx have more recent versions. While newer versions of these Office products can open an older file, an old version can’t open new versions of these file extensions.
Saving files as PDF ensures that anyone can open file extensions since you don’t need to buy Adobe Acrobat to open PDF files.
Order of precedence for user agents
Only one group is valid for a particular crawler. Google’s crawlers determine the correct group of rules by finding in the robots.txt file the group with the most specific user agent that matches the crawler’s user agent. Other groups are ignored. All non-matching text is ignored (for example, both
googlebot* are equivalent to
googlebot). The order of the groups within the robots.txt file is irrelevant.
If there’s more than one specific group declared for a user agent, all the rules from the groups applicable to the specific user agent are combined internally into a single group. User agent specific groups and global groups (
*) are not combined.
Matching of user-agent fields
This is how the crawlers would choose the relevant group:
|Group followed per crawler|
|Googlebot News|| |
|Googlebot (web)|| |
|Googlebot Images|| |
|Googlebot News (when crawling images)|| When crawling images, |
|Otherbot (web)||Other Google crawlers follow group 2.|
|Otherbot (news)|| Other Google crawlers that crawl news content, but don’t identify as |
Grouping of rules
If there are multiple groups in a robots.txt file that are relevant to a specific user agent, Google’s crawlers internally merge the groups. For example:
The crawlers internally group the rules based on user agent, for example:
Rules other than
user-agent are ignored by the robots.txt parser. This means that the following robots.txt snippet is treated as one group, and thus both
b are affected by the
disallow: / rule:
When the crawlers process the robots.txt rules, they ignore the
sitemap line. For example, this is how the crawlers would understand the previous robots.txt snippet:
- Google Drive not Uploading from Phone Error – How to Fix
- How to Upload Audio Files to Google Drive
- Google Drive Error List Retry All – How to Fix
- Failed Server Error Google Drive – How to Fix
- How to Create Subfolders in Google Drive