This module introduces key fundamentals that must be mastered to be successful in information security. Understanding web requests is essential for understanding how web applications work, which is necessary before attempting to attack or secure any web application. This makes this module the very first step in web application penetration testing.
This module will deliver these concepts through two main tools:
cURL and the
Browser DevTools. These tools are among the essential tools in any web penetration tester's arsenal, and this module will start you on the path to mastering them.
In addition to the above, this module will cover:
- An overview of the HyperText Transfer Protocol (HTTP)
- An overview of the Hypertext Transfer Protocol Secure (HTTPS)
- HTTP requests and responses and their headers
- HTTP methods and response codes
- Common HTTP methods such as GET, POST, PUT, and DELETE
- Interacting with APIs
CREST CPSA/CRT-related Sections:
- All sections
This module is broken down into sections with accompanying hands-on exercises to practice each of the tactics and techniques we cover. The module ends with a practical hands-on skills assessment to gauge your understanding of the various topic areas.
You can start and stop the module at any time and pick up where you left off. There is no time limit or "grading," but you must complete all of the exercises and the skills assessment to receive the maximum number of cubes and have this module marked as complete in any paths you have chosen.
As you work through the module, you will see example commands and command output for the various topics introduced. It is worth reproducing as many of these examples as possible to reinforce further the concepts presented in each section. You can do this in the
PwnBox provided in the interactive sections or in your virtual machine.
The module is classified as "
Fundamental" and assumes a working knowledge of the Linux command line and an understanding of information security fundamentals. Though not mandatory, we recommend taking these modules before/along with this module:
- Introduction to Networking
- Linux Fundamentals
HyperText Transfer Protocol (HTTP)
Today, the majority of the applications we use constantly interact with the internet, both web and mobile applications. Most internet communications are made with web requests through the HTTP protocol. HTTP is an application-level protocol used to access the World Wide Web resources. The term
hypertext stands for text containing links to other resources and text that the readers can easily interpret.
HTTP communication consists of a client and a server, where the client requests the server for a resource. The server processes the requests and returns the requested resource. The default port for HTTP communication is port
80, though this can be changed to any other port, depending on the web server configuration. The same requests are utilized when we use the internet to visit different websites. We enter a
Fully Qualified Domain Name (
FQDN) as a
Uniform Resource Locator (
URL) to reach the desired website, like www.hackthebox.com.
Resources over HTTP are accessed via a
URL, which offers many more specifications than simply specifying a website we want to visit. Let's look at the structure of a URL:
Here is what each component stands for:
||This is used to identify the protocol being accessed by the client, and ends with a colon and a double slash (
||This is an optional component that contains the credentials (separated by a colon
||The host signifies the resource location. This can be a hostname or an IP address|
||This points to the resource being accessed, which can be a file or a folder. If there is no path specified, the server returns the default index (e.g.
||The query string starts with a question mark (
||Fragments are processed by the browsers on the client-side to locate sections within the primary resource (e.g. a header or section on the page).|
Not all components are required to access a resource. The main mandatory fields are the scheme and the host, without which the request would have no resource to request.
The diagram above presents the anatomy of an HTTP request at a very high level. The first time a user enters the URL (
inlanefreight.com) into the browser, it sends a request to a DNS (Domain Name Resolution) server to resolve the domain and get its IP. The DNS server looks up the IP address for
inlanefreight.com and returns it. All domain names need to be resolved this way, as a server can't communicate without an IP address.
Note: Our browsers usually first look up records in the local '
/etc/hosts' file, and if the requested domain does not exist within it, then they would contact other DNS servers. We can use the '
/etc/hosts' to manually add records to for DNS resolution, by adding the IP followed by the domain name.
Once the browser gets the IP address linked to the requested domain, it sends a GET request to the default HTTP port (e.g.
80), asking for the root
/ path. Then, the web server receives the request and processes it. By default, servers are configured to return an index file when a request for
/ is received.
In this case, the contents of
index.html are read and returned by the web server as an HTTP response. The response also contains the status code (e.g.
200 OK), which indicates that the request was successfully processed. The web browser then renders the
index.html contents and presents it to the user.
Note: This module is mainly focused on HTTP web requests. For more on HTML and web applications, you may refer to the Introduction to Web Applications module.
In this module, we will be sending web requests through two of the most important tools for any web penetration tester, a Web Browser, like Chrome or Firefox, and the
cURL command line tool.
cURL (client URL) is a command-line tool and library that primarily supports HTTP along with many other protocols. This makes it a good candidate for scripts as well as automation, making it essential for sending various types of web requests from the command line, which is necessary for many types of web penetration tests.
We can send a basic HTTP request to any URL by using it as an argument for cURL, as follows:
[!bash!]$ curl inlanefreight.com <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> <html><head> ...SNIP...
We may also use cURL to download a page or a file and output the content into a file using the
-O flag. If we want to specify the output file name, we can use the
-o flag and specify the name. Otherwise, we can use
-O and cURL will use the remote file name, as follows:
[!bash!]$ curl -O inlanefreight.com/index.html % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 464 0 464 0 0 17858 0 --:--:-- --:--:-- --:--:-- 18069 [!bash!]$ ls index.html
As we can see, the output was not printed this time but rather saved into
index.html. We noticed that cURL still printed some status while processing the request. We can silent the status with the
-s flag, as follows:
[!bash!]$ curl -s -O inlanefreight.com/index.html
This time, cURL did not print anything, as the output was saved into the
index.html file. Finally, we may use the
-h flag to see what other options we may use with cURL:
[!bash!]$ curl -h Usage: curl [options...] <url> -d, --data <data> HTTP POST data -h, --help <category> Get help for commands -i, --include Include protocol response headers in the output -o, --output <file> Write to file instead of stdout -O, --remote-name Write output to a file named as the remote file -s, --silent Silent mode -u, --user <user:password> Server user and password -A, --user-agent <name> Send User-Agent <name> to server -v, --verbose Make the operation more talkative This is not the full help, this menu is stripped into categories. Use "--help category" to get an overview of all categories. Use the user manual `man curl` or the "--help all" flag for all options.
As the above message mentions, we may use
--help all to print a more detailed help menu, or
--help category (e.g.
-h http) to print the detailed help of a specific flag. If we ever need to read more detailed documentation, we can use
man curl to view the full cURL manual page.
In the upcoming sections, we will cover most of the above flags and see where we should use each of them.