Cara menggunakan get directory name javascript

How to get the path of current script using Node.js ?

Improve Article

Save Article

Like Article

  • Last Updated : 31 Jan, 2020

  • Read
  • Discuss
  • Courses
  • Practice
  • Video
  • Improve Article

    Save Article

    We can get the path of the present script in node.js by using and module scope variables.

    • __dirname: It returns the directory name of the current module in which the current script is located.
    • __filename: It returns the file name of the current module. This is the current module file’s absolute path with symlinks (symbolic links) resolved.

    Let’s Consider the below file structure of the project:

    Cara menggunakan get directory name javascript

    Below examples illustrate the use of __dirname and __filename module scope variable in node.js:

    Example 1: Determine the path of the present script while executing app.js file.

    app.js file:




    // Node.js program to demonstrates how to

    // get the current path of script

      

    // To get the filename

    console.log(`Filename is ${__filename}`);

      

    // To get the directory name

    console.log(`Directory name is ${__dirname}`);

    Output:

    Filename is D:\DemoProject\app.js
    Directory name is D:\DemoProject
    

    Example 2: Determine the path of the present script while executing routes\user.js file.

    user.js file:




    // Node.js program to demonstrates how to

    // get the current path of script

      

    // To get the filename

    console.log(`Filename is ${__filename}`);

      

    // To get the directory name

    console.log(`Directory name is ${__dirname}`);

    Output:

    Filename is D:\DemoProject\routes\app.js
    Directory name is D:\DemoProject\routes
    

    My Personal Notes arrow_drop_up

    Save

    Please Login to comment...

    If you use a site hosting service, such as Wix or Blogger, you might not need to (or be able to) edit your robots.txt file directly. Instead, your provider might expose a search settings page or some other mechanism to tell search engines whether or not to crawl your page.

    If you want to hide or unhide one of your pages from search engines, search for instructions about modifying your page visibility in search engines on your hosting service, for example, search for "wix hide page from search engines".

    You can control which files crawlers may access on your site with a robots.txt file.

    A robots.txt file lives at the root of your site. So, for site

    # Example 1: Block only Googlebot
    User-agent: Googlebot
    Disallow: /
    
    # Example 2: Block Googlebot and Adsbot
    User-agent: Googlebot
    User-agent: AdsBot-Google
    Disallow: /
    
    # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
    User-agent: *
    Disallow: /
    4, the robots.txt file lives at
    # Example 1: Block only Googlebot
    User-agent: Googlebot
    Disallow: /
    
    # Example 2: Block Googlebot and Adsbot
    User-agent: Googlebot
    User-agent: AdsBot-Google
    Disallow: /
    
    # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
    User-agent: *
    Disallow: /
    5. robots.txt is a plain text file that follows the . A robots.txt file consists of one or more rules. Each rule blocks or allows access for all or a specific crawler to a specified file path on the domain or subdomain where the robots.txt file is hosted. Unless you specify otherwise in your robots.txt file, all files are implicitly allowed for crawling.

    Here is a simple robots.txt file with two rules:

    User-agent: Googlebot
    Disallow: /nogooglebot/
    
    User-agent: *
    Allow: /
    
    Sitemap: https://www.example.com/sitemap.xml

    Here's what that robots.txt file means:

    1. The user agent named Googlebot is not allowed to crawl any URL that starts with
      # Example 1: Block only Googlebot
      User-agent: Googlebot
      Disallow: /
      
      # Example 2: Block Googlebot and Adsbot
      User-agent: Googlebot
      User-agent: AdsBot-Google
      Disallow: /
      
      # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
      User-agent: *
      Disallow: /
      6.
    2. All other user agents are allowed to crawl the entire site. This could have been omitted and the result would be the same; the default behavior is that user agents are allowed to crawl the entire site.
    3. The site's sitemap file is located at
      # Example 1: Block only Googlebot
      User-agent: Googlebot
      Disallow: /
      
      # Example 2: Block Googlebot and Adsbot
      User-agent: Googlebot
      User-agent: AdsBot-Google
      Disallow: /
      
      # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
      User-agent: *
      Disallow: /
      7.

    See the section for more examples.

    Basic guidelines for creating a robots.txt file

    Creating a robots.txt file and making it generally accessible and useful involves four steps:

    1. .
    2. .
    3. .
    4. .

    Create a robots.txt file

    You can use almost any text editor to create a robots.txt file. For example, Notepad, TextEdit, vi, and emacs can create valid robots.txt files. Don't use a word processor; word processors often save files in a proprietary format and can add unexpected characters, such as curly quotes, which can cause problems for crawlers. Make sure to save the file with UTF-8 encoding if prompted during the save file dialog.

    Format and location rules:

    • The file must be named robots.txt.
    • Your site can have only one robots.txt file.
    • The robots.txt file must be located at the root of the website host to which it applies. For instance, to control crawling on all URLs below
      # Example 1: Block only Googlebot
      User-agent: Googlebot
      Disallow: /
      
      # Example 2: Block Googlebot and Adsbot
      User-agent: Googlebot
      User-agent: AdsBot-Google
      Disallow: /
      
      # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
      User-agent: *
      Disallow: /
      8, the robots.txt file must be located at
      # Example 1: Block only Googlebot
      User-agent: Googlebot
      Disallow: /
      
      # Example 2: Block Googlebot and Adsbot
      User-agent: Googlebot
      User-agent: AdsBot-Google
      Disallow: /
      
      # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
      User-agent: *
      Disallow: /
      9. It cannot be placed in a subdirectory (for example, at
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      0). If you're unsure about how to access your website root, or need permissions to do so, contact your web hosting service provider. If you can't access your website root, use an alternative blocking method such as meta tags.
    • A robots.txt file can be posted on a subdomain (for example,
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      1) or on non-standard ports (for example,
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      2).
    • A robots.txt file applies only to paths within the protocol, host, and port where it is posted. That is, rules in
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      3 apply only to files in
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      4, not to subdomains such as
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      5, or alternate protocols, such as
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      6.
    • A robots.txt file must be an UTF-8 encoded text file (which includes ASCII). Google may ignore characters that are not part of the UTF-8 range, potentially rendering robots.txt rules invalid.

    How to write robots.txt rules

    Rules are instructions for crawlers about which parts of your site they can crawl. Follow these guidelines when adding rules to your robots.txt file:

    • A robots.txt file consists of one or more groups (set of rules).
    • Each group consists of multiple rules (also known as directives), one rule per line. Each group begins with a
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      7 line that specifies the target of the groups.
    • A group gives the following information:
      • Who the group applies to (the user agent).
      • Which directories or files that agent can access.
      • Which directories or files that agent cannot access.
    • Crawlers process groups from top to bottom. A user agent can match only one rule set, which is the first, most specific group that matches a given user agent. If there are multiple groups for the same user agent, the groups will be combined into a single group before processing.
    • The default assumption is that a user agent can crawl any page or directory not blocked by a
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      8 rule.
    • Rules are case-sensitive. For instance,
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      9 applies to
      User-agent: *
      Disallow: /
      0, but not
      User-agent: *
      Disallow: /
      1.
    • The
      User-agent: *
      Disallow: /
      2 character marks the beginning of a comment. Comments are ignored during processing.

    Google's crawlers support the following rules in robots.txt files:

    • User-agent: *
      Disallow: /
      3 [Required, one or more per group] The rule specifies the name of the automatic client known as search engine crawler that the rule applies to. This is the first line for any rule group. Google user agent names are listed in the Google list of user agents. Using an asterisk (
      User-agent: *
      Disallow: /
      4) matches all crawlers except the various AdsBot crawlers, which must be named explicitly. For example:
      # Example 1: Block only Googlebot
      User-agent: Googlebot
      Disallow: /
      
      # Example 2: Block Googlebot and Adsbot
      User-agent: Googlebot
      User-agent: AdsBot-Google
      Disallow: /
      
      # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
      User-agent: *
      Disallow: /
    • User-agent: *
      Disallow: /
      5 [At least one or more
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      8 or
      User-agent: *
      Disallow: /
      7 entries per rule] A directory or page, relative to the root domain, that you don't want the user agent to crawl. If the rule refers to a page, it must be the full page name as shown in the browser. It must start with a
      User-agent: *
      Disallow: /
      8 character and if it refers to a directory, it must end with the
      User-agent: *
      Disallow: /
      8 mark.
    • User-agent: *
      Disallow: /calendar/
      Disallow: /junk/
      Disallow: /books/fiction/contemporary/
      0 [At least one or more
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      8 or
      User-agent: *
      Disallow: /
      7 entries per rule] A directory or page, relative to the root domain, that may be crawled by the user agent just mentioned. This is used to override a
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml
      8 rule to allow crawling of a subdirectory or page in a disallowed directory. For a single page, specify the full page name as shown in the browser. It must start with a
      User-agent: *
      Disallow: /
      8 character and if it refers to a directory, it must end with the
      User-agent: *
      Disallow: /
      8 mark.
    • User-agent: *
      Disallow: /calendar/
      Disallow: /junk/
      Disallow: /books/fiction/contemporary/
      6 [Optional, zero or more per file] The location of a sitemap for this website. The sitemap URL must be a fully-qualified URL; Google doesn't assume or check http/https/www.non-www alternates. Sitemaps are a good way to indicate which content Google should crawl, as opposed to which content it can or cannot crawl. Learn more about sitemaps. Example:
      Sitemap: https://example.com/sitemap.xml
      Sitemap: https://www.example.com/sitemap.xml

    All rules, except

    User-agent: *
    Disallow: /calendar/
    Disallow: /junk/
    Disallow: /books/fiction/contemporary/
    7, support the
    User-agent: *
    Disallow: /
    4 wildcard for a path prefix, suffix, or entire string.

    Lines that don't match any of these rules are ignored.

    Read our page about Google's interpretation of the robots.txt specification for the complete description of each rule.

    Upload the robots.txt file

    Once you saved your robots.txt file to your computer, you're ready to make it available to search engine crawlers. There's no one tool that can help you with this, because how you upload the robots.txt file to your site depends on your site and server architecture. Get in touch with your hosting company or search the documentation of your hosting company; for example, search for "upload files infomaniak".

    After you upload the robots.txt file, test whether it's publicly accessible and if Google can parse it.

    Test robots.txt markup

    To test whether your newly uploaded robots.txt file is publicly accessible, open a private browsing window (or equivalent) in your browser and navigate to the location of the robots.txt file. For example,

    Sitemap: https://example.com/sitemap.xml
    Sitemap: https://www.example.com/sitemap.xml
    3. If you see the contents of your robots.txt file, you're ready to test the markup.

    Google offers two options for testing robots.txt markup:

    1. The robots.txt Tester in Search Console. You can only use this tool for robots.txt files that are already accessible on your site.
    2. If you're a developer, check out and build Google's open source robots.txt library, which is also used in Google Search. You can use this tool to test robots.txt files locally on your computer.

    Submit robots.txt file to Google

    Once you uploaded and tested your robots.txt file, Google's crawlers will automatically find and start using your robots.txt file. You don't have to do anything. If you updated your robots.txt file and you need to refresh Google's cached copy as soon as possible, learn how to submit an updated robots.txt file.

    Useful robots.txt rules

    Here are some common useful robots.txt rules:

    Useful rulesDisallow crawling of the entire website

    Keep in mind that in some situations URLs from the website may still be indexed, even if they haven't been crawled.

    Note: This does not match the various AdsBot crawlers, which must be named explicitly.
    User-agent: *
    Disallow: /
    Disallow crawling of a directory and its contents

    Append a forward slash to the directory name to disallow crawling of a whole directory.

    Caution: Remember, don't use robots.txt to block access to private content; use proper authentication instead. URLs disallowed by the robots.txt file might still be indexed without being crawled, and the robots.txt file can be viewed by anyone, potentially disclosing the location of your private content.
    User-agent: *
    Disallow: /calendar/
    Disallow: /junk/
    Disallow: /books/fiction/contemporary/
    Allow access to a single crawler

    Only

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    0 may crawl the whole site.

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    Allow access to all but a single crawler

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    1 may not crawl the site, all other bots may.

    User-agent: Unnecessarybot
    Disallow: /
    
    User-agent: *
    Allow: /

    Disallow crawling of a single web page

    For example, disallow the

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    2 page located at
    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    3, and
    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    4 in the
    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    5 directory.

    User-agent: *
    Disallow: /useless_file.html
    Disallow: /junk/other_useless_file.html

    Disallow crawling of the whole site except a subdirectory

    Crawlers may only access the

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    6 subdirectory.

    User-agent: *
    Disallow: /
    Allow: /public/

    Block a specific image from Google Images

    For example, disallow the

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    7 image.

    User-agent: Googlebot-Image
    Disallow: /images/dogs.jpg

    Block all images on your site from Google Images

    Google can't index images and videos without crawling them.

    # Example 1: Block only Googlebot
    User-agent: Googlebot
    Disallow: /
    
    # Example 2: Block Googlebot and Adsbot
    User-agent: Googlebot
    User-agent: AdsBot-Google
    Disallow: /
    
    # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
    User-agent: *
    Disallow: /
    0

    Disallow crawling of files of a specific file type

    For example, disallow for crawling all

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    8 files.

    # Example 1: Block only Googlebot
    User-agent: Googlebot
    Disallow: /
    
    # Example 2: Block Googlebot and Adsbot
    User-agent: Googlebot
    User-agent: AdsBot-Google
    Disallow: /
    
    # Example 3: Block all crawlers except AdsBot (AdsBot crawlers must be named explicitly)
    User-agent: *
    Disallow: /
    1

    Disallow crawling of an entire site, but allow

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    9

    This implementation hides your pages from search results, but the

    User-agent: Googlebot-news
    Allow: /
    
    User-agent: *
    Disallow: /
    9 web crawler can still analyze them to decide what ads to show visitors on your site.